Bird'S Eye View
Bird's-Eye-View (BEV) representation transforms multi-camera images into a top-down view, crucial for autonomous driving and robotics by providing a unified, geometrically-structured scene understanding. Current research focuses on improving BEV generation accuracy and robustness using transformer-based architectures, often incorporating multimodal sensor fusion (camera, LiDAR, radar) and advanced techniques like masked attention and Gaussian splatting to enhance feature representation and handle challenges like occlusion and domain adaptation. This work is significant for advancing autonomous systems by enabling more reliable perception, particularly in complex or challenging environments, and improving the performance of downstream tasks such as object detection, mapping, and trajectory prediction.
Papers
QuadBEV: An Efficient Quadruple-Task Perception Framework via Bird's-Eye-View Representation
Yuxin Li, Yiheng Li, Xulei Yang, Mengying Yu, Zihang Huang, Xiaojun Wu, Chai Kiat Yeo
Learning Content-Aware Multi-Modal Joint Input Pruning via Bird's-Eye-View Representation
Yuxin Li, Yiheng Li, Xulei Yang, Mengying Yu, Zihang Huang, Xiaojun Wu, Chai Kiat Yeo
Mask2Map: Vectorized HD Map Construction Using Bird's Eye View Segmentation Masks
Sehwan Choi, Jungho Kim, Hongjae Shin, Jun Won Choi
OE-BevSeg: An Object Informed and Environment Aware Multimodal Framework for Bird's-eye-view Vehicle Semantic Segmentation
Jian Sun, Yuqi Dai, Chi-Man Vong, Qing Xu, Shengbo Eben Li, Jianqiang Wang, Lei He, Keqiang Li