SkillAgentSearch skills...

VideoCanvas

Official Code of "VideoCanvas: Unified Video Completion from Arbitrary Spatiotemporal Patches via In-Context Conditioning"

Install / Use

/learn @KlingAIResearch/VideoCanvas
About this skill

Quality Score

0/100

Supported Platforms

Universal

README

VideoCanvas: Unified Video Completion from Arbitrary Spatiotemporal Patches via In-Context Conditioning

Minghong Cai<sup>1 †</sup>, Qiulin Wang<sup>2 ✉</sup>, Zongli Ye<sup>1</sup>, Wenze Liu<sup>1</sup>, Quande Liu<sup>2</sup>, Weicai Ye<sup>2</sup>, Xintao Wang<sup>2</sup>, Pengfei Wan<sup>2</sup>, Kun Gai<sup>2</sup>, Xiangyu Yue<sup>1 ✉</sup> <br> <sup>1</sup>MMLab, The Chinese University of Hong Kong <sup>2</sup>Kling Team, Kuaishou Technology <br> †: Intern at Kuaishou Technology, ✉: Corresponding Authors

<a href='https://arxiv.org/abs/2510.08555'><img src='https://img.shields.io/badge/ArXiv-2510.08555-red'></a> <a href='https://onevfall.github.io/project_page/videocanvas/#'><img src='https://img.shields.io/badge/Project-Page-Green'></a>

📋 News

  • [2025.10.9] Release Arxiv paper.

📖 Introduction

VideoCanvas has two key contributions:

  • 🎯 Unified Tasks: VideoCanvas introduces a unified paradigm for arbitrary spatio-temporal video generation, seamlessly integrating diverse capabilities including image/patch-to-video conditioning at any timestamp, inpainting/outpainting, camera control, scene transitions, and video extension.
  • 🛠️ Simple Solution: Our technical innovation leverages In-Context Conditioning with zero-padding for spatial control and Temporal RoPE Interpolation for temporal alignment, achieving frame-precise video generation without fine-tuning VAEs or adding parameters. <br>

https://github.com/user-attachments/assets/c3777c16-0a1e-4cbb-af6f-856f88312317

📖 VideoCanvasBench

We will release this benchmark, including intra-scene and inter-scene evaluation data.

⚙️ Code (Coming soon)

Citation

 @article{cai2025videocanvas,
    title={VideoCanvas: Unified Video Completion from Arbitrary Spatiotemporal Patches via In-Context Conditioning},
    author={Minghong Cai, Qiulin Wang, Zongli Ye, Wenze Liu, Quande Liu, Weicai Ye, Xintao Wang, Pengfei Wan, Kun Gai, Xiangyu Yue},
    journal={arXiv preprint arXiv:2510.08555},
    year={2025}
}
View on GitHub
GitHub Stars71
CategoryContent
Updated3d ago
Forks0

Security Score

80/100

Audited on Mar 28, 2026

No findings