Bird'S Eye View Map Segmentation
Bird's-eye-view (BEV) map segmentation aims to create a top-down representation of a scene, crucial for autonomous driving and other applications requiring comprehensive spatial understanding. Current research emphasizes developing robust and efficient models, often employing transformer-based architectures and multi-modal sensor fusion (e.g., combining camera and LiDAR data) to improve accuracy and generalization across diverse environments. A key focus is enhancing model robustness to sensor failures and variations in data, improving the reliability of BEV representations for tasks like 3D object detection and scene understanding. This work has significant implications for advancing autonomous driving safety and efficiency.
Papers
CoBEVT: Cooperative Bird's Eye View Semantic Segmentation with Sparse Transformers
Runsheng Xu, Zhengzhong Tu, Hao Xiang, Wei Shao, Bolei Zhou, Jiaqi Ma
Vision-based Uneven BEV Representation Learning with Polar Rasterization and Surface Estimation
Zhi Liu, Shaoyu Chen, Xiaojie Guo, Xinggang Wang, Tianheng Cheng, Hongmei Zhu, Qian Zhang, Wenyu Liu, Yi Zhang