Camera Control
Camera control in video generation and manipulation is a rapidly advancing field focused on achieving precise and intuitive camera movement through various input methods, such as text prompts or user-defined trajectories. Current research emphasizes the integration of deep learning models, including diffusion models and transformers, often incorporating techniques like ControlNet and attention mechanisms to enable fine-grained control over camera pose and motion, even in multi-view and multi-video scenarios. This work has significant implications for fields like autonomous driving (through improved simulation), film production (via automated directing and cinematography), and 3D scene modeling (with enhanced realism and user control).
Papers
Boosting Camera Motion Control for Video Diffusion Transformers
Soon Yau Cheong, Duygu Ceylan, Armin Mustafa, Andrew Gilbert, Chun-Hao Paul Huang
Cavia: Camera-controllable Multi-view Video Diffusion with View-Integrated Attention
Dejia Xu, Yifan Jiang, Chen Huang, Liangchen Song, Thorsten Gernoth, Liangliang Cao, Zhangyang Wang, Hao Tang