Bird'S Eye View
Bird's-Eye-View (BEV) representation transforms multi-camera images into a top-down view, crucial for autonomous driving and robotics by providing a unified, geometrically-structured scene understanding. Current research focuses on improving BEV generation accuracy and robustness using transformer-based architectures, often incorporating multimodal sensor fusion (camera, LiDAR, radar) and advanced techniques like masked attention and Gaussian splatting to enhance feature representation and handle challenges like occlusion and domain adaptation. This work is significant for advancing autonomous systems by enabling more reliable perception, particularly in complex or challenging environments, and improving the performance of downstream tasks such as object detection, mapping, and trajectory prediction.
Papers
Mask2Map: Vectorized HD Map Construction Using Bird's Eye View Segmentation Masks
Sehwan Choi, Jungho Kim, Hongjae Shin, Jun Won Choi
OE-BevSeg: An Object Informed and Environment Aware Multimodal Framework for Bird's-eye-view Vehicle Semantic Segmentation
Jian Sun, Yuqi Dai, Chi-Man Vong, Qing Xu, Shengbo Eben Li, Jianqiang Wang, Lei He, Keqiang Li
Map It Anywhere (MIA): Empowering Bird's Eye View Mapping using Large-scale Public Data
Cherie Ho, Jiaye Zou, Omar Alama, Sai Mitheran Jagadesh Kumar, Benjamin Chiang, Taneesh Gupta, Chen Wang, Nikhil Keetha, Katia Sycara, Sebastian Scherer
BLOS-BEV: Navigation Map Enhanced Lane Segmentation Network, Beyond Line of Sight
Hang Wu, Zhenghao Zhang, Siyuan Lin, Tong Qin, Jin Pan, Qiang Zhao, Chunjing Xu, Ming Yang