Video Dynamic
Video dynamic research focuses on understanding and manipulating the temporal evolution of visual information in videos. Current efforts concentrate on improving video generation and editing through techniques like diffusion models, neural ordinary differential equations, and attention mechanisms that explicitly model temporal relationships, often incorporating cross-modal information from audio or text. These advancements are driving progress in applications such as video question answering, audio-visual speech recognition, and high-quality video editing, impacting fields ranging from computer vision to media analysis.
Papers
NaRCan: Natural Refined Canonical Image with Integration of Diffusion Prior for Video Editing
Ting-Hsuan Chen, Jiewen Chan, Hau-Shiang Shiu, Shih-Han Yen, Chang-Han Yeh, Yu-Lun Liu
AID: Adapting Image2Video Diffusion Models for Instruction-guided Video Prediction
Zhen Xing, Qi Dai, Zejia Weng, Zuxuan Wu, Yu-Gang Jiang
HDR-ChipQA: No-Reference Quality Assessment on High Dynamic Range Videos
Joshua P. Ebenezer, Zaixi Shang, Yongjun Wu, Hai Wei, Sriram Sethuraman, Alan C. Bovik
Making Video Quality Assessment Models Robust to Bit Depth
Joshua P. Ebenezer, Zaixi Shang, Yongjun Wu, Hai Wei, Sriram Sethuraman, Alan C. Bovik