Motion Representation
Motion representation in computer vision and robotics focuses on efficiently and accurately encoding movement information from various sources, such as videos, sensor data, and textual descriptions, to enable tasks like motion prediction, generation, and tracking. Current research emphasizes developing robust and generalizable representations using techniques like transformers, diffusion models, and variational autoencoders, often incorporating self-supervised learning and contrastive methods to improve performance. These advancements are crucial for improving human-computer interaction, autonomous systems, and animation technologies by enabling more realistic and nuanced modeling of movement.
Papers
Motion Graph Unleashed: A Novel Approach to Video Prediction
Yiqi Zhong, Luming Liang, Bohan Tang, Ilya Zharkov, Ulrich Neumann
MotionGPT-2: A General-Purpose Motion-Language Model for Motion Generation and Understanding
Yuan Wang, Di Huang, Yaqi Zhang, Wanli Ouyang, Jile Jiao, Xuetao Feng, Yan Zhou, Pengfei Wan, Shixiang Tang, Dan Xu