Video Generation Model
Video generation models aim to create realistic and coherent video sequences from various inputs, such as text descriptions, images, or other videos, focusing on improving visual fidelity, temporal consistency, and user control. Current research heavily utilizes diffusion models, often incorporating techniques like attention mechanisms, multi-agent frameworks, and noise rescheduling to enhance generation quality and efficiency, addressing challenges like long video generation and multi-scene consistency. These advancements have significant implications for diverse fields, including film production, robotics, medical simulation, and the creation of more realistic and interactive digital content.
Papers
EVA: An Embodied World Model for Future Video Anticipation
Xiaowei Chi, Hengyuan Zhang, Chun-Kai Fan, Xingqun Qi, Rongyu Zhang, Anthony Chen, Chi-min Chan, Wei Xue, Wenhan Luo, Shanghang Zhang, Yike Guo
Allegro: Open the Black Box of Commercial-Level Video Generation Model
Yuan Zhou, Qiuyue Wang, Yuxuan Cai, Huan Yang
T2V-Turbo-v2: Enhancing Video Generation Model Post-Training through Data, Reward, and Conditional Guidance Design
Jiachen Li, Qian Long, Jian Zheng, Xiaofeng Gao, Robinson Piramuthu, Wenhu Chen, William Yang Wang
TweedieMix: Improving Multi-Concept Fusion for Diffusion-based Image/Video Generation
Gihyun Kwon, Jong Chul Ye
Gen2Act: Human Video Generation in Novel Scenarios enables Generalizable Robot Manipulation
Homanga Bharadhwaj, Debidatta Dwibedi, Abhinav Gupta, Shubham Tulsiani, Carl Doersch, Ted Xiao, Dhruv Shah, Fei Xia, Dorsa Sadigh, Sean Kirmani
Technical Report: Competition Solution For Modelscope-Sora
Shengfu Chen, Hailong Liu, Wenzhao Wei
Training-free Long Video Generation with Chain of Diffusion Model Experts
Wenhao Li, Yichao Cao, Xiu Su, Xi Lin, Shan You, Mingkai Zheng, Yi Chen, Chang Xu
TVG: A Training-free Transition Video Generation Method with Diffusion Models
Rui Zhang, Yaosen Chen, Yuegen Liu, Wei Wang, Xuming Wen, Hongxia Wang