3D Detection
3D object detection aims to accurately identify and locate objects in three-dimensional space from various sensor inputs, primarily for applications like autonomous driving and robotics. Current research emphasizes improving robustness and generalization across diverse datasets and challenging conditions (e.g., varying weather, occlusions) using techniques like multi-dataset training, temporal information fusion, and diffusion models. Prominent approaches involve transformer-based architectures, BEV (bird's-eye-view) transformations, and innovative data augmentation strategies to address data scarcity and annotation costs. Advancements in this field are crucial for enhancing the safety and reliability of autonomous systems and other applications requiring precise 3D scene understanding.
Papers
MV-JAR: Masked Voxel Jigsaw and Reconstruction for LiDAR-Based Self-Supervised Pre-Training
Runsen Xu, Tai Wang, Wenwei Zhang, Runjian Chen, Jinkun Cao, Jiangmiao Pang, Dahua Lin
Collaboration Helps Camera Overtake LiDAR in 3D Detection
Yue Hu, Yifan Lu, Runsheng Xu, Weidi Xie, Siheng Chen, Yanfeng Wang
LoGoNet: Towards Accurate 3D Object Detection with Local-to-Global Cross-Modal Fusion
Xin Li, Tao Ma, Yuenan Hou, Botian Shi, Yuchen Yang, Youquan Liu, Xingjiao Wu, Qin Chen, Yikang Li, Yu Qiao, Liang He
Calibration-free BEV Representation for Infrastructure Perception
Siqi Fan, Zhe Wang, Xiaoliang Huo, Yan Wang, Jingjing Liu