Single Video
Single-video analysis focuses on reconstructing detailed 3D models and animations from a single video input, overcoming the limitations of traditional methods requiring multiple views. Current research emphasizes novel neural network architectures, including neural radiance fields (NeRFs) and transformers, to achieve high-fidelity 4D reconstructions of dynamic scenes, articulated objects, and even human avatars, often incorporating techniques like test-time training and implicit representations. This field is significant for its potential applications in virtual and augmented reality, robotics (e.g., imitation learning), and computer graphics, enabling the creation of realistic digital environments and avatars from readily available video data.
Papers
HOSNeRF: Dynamic Human-Object-Scene Neural Radiance Fields from a Single Video
Jia-Wei Liu, Yan-Pei Cao, Tianyuan Yang, Eric Zhongcong Xu, Jussi Keppo, Ying Shan, Xiaohu Qie, Mike Zheng Shou
Efficient Robot Skill Learning with Imitation from a Single Video for Contact-Rich Fabric Manipulation
Shengzeng Huo, Anqing Duan, Lijun Han, Luyin Hu, Hesheng Wang, David Navarro-Alarcon