3D Semantic Scene
3D semantic scene understanding aims to create comprehensive, computer-interpretable models of real-world environments, capturing both the geometry and semantic labels of objects and spaces. Current research emphasizes improving the accuracy and efficiency of scene reconstruction from various data sources (LiDAR, RGB-D cameras, multi-view videos), often employing deep learning architectures like neural radiance fields (NeRFs) and transformers, along with techniques such as contrastive learning and multi-scan fusion to handle challenges like long distances and occlusions. This field is crucial for advancing robotics, autonomous navigation, augmented reality, and other applications requiring robust scene interpretation and interaction.
Papers
Frankenstein: Generating Semantic-Compositional 3D Scenes in One Tri-Plane
Han Yan, Yang Li, Zhennan Wu, Shenzhou Chen, Weixuan Sun, Taizhang Shang, Weizhe Liu, Tian Chen, Xiaqiang Dai, Chao Ma, Hongdong Li, Pan Ji
Semantic Is Enough: Only Semantic Information For NeRF Reconstruction
Ruibo Wang, Song Zhang, Ping Huang, Donghai Zhang, Wei Yan