LongLive
[ICLR 2026] LongLive: Real-time Interactive Long Video Generation
Install / Use
/learn @NVlabs/LongLiveREADME
🎬 LongLive: Real-time Interactive Long Video Generation
<div align="center"> </div>💡 TLDR: Turn interactive prompts into long videos—instantly, as you type!
LongLive: Real-time Interactive Long Video Generation [Paper] <br /> Shuai Yang, Wei Huang, Ruihang Chu, Yicheng Xiao, Yuyang Zhao, Xianbang Wang, Muyang Li, Enze Xie, Yingcong Chen, Yao Lu, Song Han, Yukang Chen <br />
TABLE OF CONTENTS
News
- [x] [2026.1.27] LongLive is accepted by ICLR-2026. 🎉🎉🎉
- [x] [2026.1.11] Many thanks @qixinhu11 for adapting LongLive's original RoPE into KV-cache relative RoPE. Now LongLive supports generating infinite long videos!
- [x] [2025.12.4] We fix a bug in
global_sink==Falsemode. Now our model generate videos in higher quality. - [x] [2025.11.3] We implement LongLive on linear attention model SANA-Video! Now SANA-Video can generate 60s interactive videos in real-time.
- [x] [2025.11.1] The license has been changed from CC-BY-NC-SA 4.0 to Apache 2.0.
- [x] [2025.10.11] Many thanks to @yondonfu for building an interactive UI based on LongLive. Please check it here.
- [x] [2025.10.1] We compare Sora2 (+ GPT-5 prompt engineering) with LongLive-1.3B in the interactive long video generation. See here for details.
- [x] [2025.9.30] We release example prompts to reproduce our demo videos.
- [x] [2025.9.29] We release Paper, this GitHub repo LongLive with all training and inference code, the model weight LongLive-1.3B, and demo page Website.
Highlights
- Long Video Gen: LongLive supports up to 240s video generation, with visual consistency.
- Real-time Inference: LongLive supports 20.7 FPS generation speed on a single H100 GPU, and 24.8 FPS with FP8 quantization with marginal quality loss.
- Efficient Fine-tuning: LongLive extends a short-clip model to minute-long generation in 32 H100 GPU-days.
Introduction
<p align="center" style="border-radius: 10px"> <img src="assets/pipeline.jpg" width="100%" alt="logo"/> <strong>LongLive accepts sequential user prompts and generates corresponding videos in real time, enabling user-guided long video generation.</strong> </p>Please see our docs for Installation, Training, and Inference.
How to contribute
- Make sure to have git installed.
- Create your own fork of the project.
- Clone the repository on your local machine, using git clone and pasting the url of this project.
- Read both the
RequirementsandInstallation and Quick Guidesections below. - Commit and push your changes.
- Make a pull request when finished modifying the project.
Citation
Please consider to cite our paper and this framework, if they are helpful in your research.
@article{yang2025longlive,
title={LongLive: Real-time Interactive Long Video Generation},
author={Shuai Yang and Wei Huang and Ruihang Chu and Yicheng Xiao and Yuyang Zhao and Xianbang Wang and Muyang Li and Enze Xie and Yingcong Chen and Yao Lu and Song Hanand Yukang Chen},
year={2025},
eprint={2509.22622},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
Acknowledgement
- Self-Forcing: the codebase and algorithm we built upon. Thanks for their wonderful work.
- Wan: the base model we built upon. Thanks for their wonderful work.
Related Skills
qqbot-channel
345.9kQQ 频道管理技能。查询频道列表、子频道、成员、发帖、公告、日程等操作。使用 qqbot_channel_api 工具代理 QQ 开放平台 HTTP 接口,自动处理 Token 鉴权。当用户需要查看频道、管理子频道、查询成员、发布帖子/公告/日程时使用。
docs-writer
100.0k`docs-writer` skill instructions As an expert technical writer and editor for the Gemini CLI project, you produce accurate, clear, and consistent documentation. When asked to write, edit, or revie
model-usage
345.9kUse CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
arscontexta
2.9kClaude Code plugin that generates individualized knowledge systems from conversation. You describe how you think and work, have a conversation and get a complete second brain as markdown files you own.

