Video Diffusion Model
Video diffusion models generate videos by iteratively removing noise from random data, guided by text prompts or other conditioning information. Current research focuses on improving temporal consistency, enhancing video quality (especially at high resolutions), and developing efficient algorithms for long video generation and various control mechanisms (e.g., camera control, object manipulation). These advancements are significant for applications in film production, animation, and 3D modeling, offering powerful tools for creating realistic and controllable video content.
Papers
Uncovering Hidden Subspaces in Video Diffusion Models Using Re-Identification
Mischa Dombrowski, Hadrien Reynaud, Bernhard Kainz
DimensionX: Create Any 3D and 4D Scenes from a Single Image with Controllable Video Diffusion
Wenqiang Sun, Shuo Chen, Fangfu Liu, Zilong Chen, Yueqi Duan, Jun Zhang, Yikai Wang
ReferEverything: Towards Segmenting Everything We Can Speak of in Videos
Anurag Bagchi, Zhipeng Bao, Yu-Xiong Wang, Pavel Tokmakov, Martial Hebert
LumiSculpt: A Consistency Lighting Control Network for Video Generation
Yuxin Zhang, Dandan Zheng, Biao Gong, Jingdong Chen, Ming Yang, Weiming Dong, Changsheng Xu