Point Cloud Video
Point cloud video (PCV) research focuses on effectively processing and understanding sequences of 3D point cloud data, aiming to extract meaningful information about dynamic scenes. Current research emphasizes efficient representation learning using architectures like state space models and transformers, often incorporating techniques such as contrastive learning and cross-modal knowledge transfer from image data to overcome challenges posed by the unstructured nature of point clouds. These advancements are crucial for applications in robotics, autonomous driving, and human-computer interaction, enabling improved scene understanding, action recognition, and anomaly detection in dynamic 3D environments. The development of more efficient and accurate models is a key focus, particularly for long video sequences.
Papers
Point Contrastive Prediction with Semantic Clustering for Self-Supervised Learning on Point Cloud Videos
Xiaoxiao Sheng, Zhiqiang Shen, Gang Xiao, Longguang Wang, Yulan Guo, Hehe Fan
Masked Spatio-Temporal Structure Prediction for Self-supervised Learning on Point Cloud Videos
Zhiqiang Shen, Xiaoxiao Sheng, Hehe Fan, Longguang Wang, Yulan Guo, Qiong Liu, Hao Wen, Xi Zhou