Video Diffusion
Video diffusion models are revolutionizing video generation by leveraging the power of diffusion processes to create high-quality, temporally consistent videos. Current research focuses on improving temporal modeling through novel architectures like vectorized timesteps and incorporating diverse control signals (e.g., sketches, depth maps) for fine-grained manipulation of video content. These advancements are significantly impacting various applications, including image-to-video generation, video editing, novel view synthesis, and the creation of physically realistic animations, pushing the boundaries of generative modeling in multimedia. The ability to generate and manipulate videos with increased control and realism has broad implications for fields ranging from entertainment and special effects to scientific visualization and virtual reality.
Papers
Contrastive Sequential-Diffusion Learning: Non-linear and Multi-Scene Instructional Video Synthesis
Vasco Ramos, Yonatan Bitton, Michal Yarom, Idan Szpektor, Joao Magalhaes
Animate3D: Animating Any 3D Model with Multi-view Video Diffusion
Yanqin Jiang, Chaohui Yu, Chenjie Cao, Fan Wang, Weiming Hu, Jin Gao
Learning Temporally Consistent Video Depth from Video Diffusion Priors
Jiahao Shao, Yuanbo Yang, Hongyu Zhou, Youmin Zhang, Yujun Shen, Matteo Poggi, Yiyi Liao
DreamPhysics: Learning Physical Properties of Dynamic 3D Gaussians with Video Diffusion Priors
Tianyu Huang, Haoze Zhang, Yihan Zeng, Zhilu Zhang, Hui Li, Wangmeng Zuo, Rynson W. H. Lau