Bird'S Eye View
Bird's-Eye-View (BEV) representation transforms multi-camera images into a top-down view, crucial for autonomous driving and robotics by providing a unified, geometrically-structured scene understanding. Current research focuses on improving BEV generation accuracy and robustness using transformer-based architectures, often incorporating multimodal sensor fusion (camera, LiDAR, radar) and advanced techniques like masked attention and Gaussian splatting to enhance feature representation and handle challenges like occlusion and domain adaptation. This work is significant for advancing autonomous systems by enabling more reliable perception, particularly in complex or challenging environments, and improving the performance of downstream tasks such as object detection, mapping, and trajectory prediction.
Papers
Map It Anywhere (MIA): Empowering Bird's Eye View Mapping using Large-scale Public Data
Cherie Ho, Jiaye Zou, Omar Alama, Sai Mitheran Jagadesh Kumar, Benjamin Chiang, Taneesh Gupta, Chen Wang, Nikhil Keetha, Katia Sycara, Sebastian Scherer
BLOS-BEV: Navigation Map Enhanced Lane Segmentation Network, Beyond Line of Sight
Hang Wu, Zhenghao Zhang, Siyuan Lin, Tong Qin, Jin Pan, Qiang Zhao, Chunjing Xu, Ming Yang
Deep Bayesian Future Fusion for Self-Supervised, High-Resolution, Off-Road Mapping
Shubhra Aich, Wenshan Wang, Parv Maheshwari, Matthew Sivaprakasam, Samuel Triest, Cherie Ho, Jason M. Gregory, John G. Rogers III, Sebastian Scherer
BEVCar: Camera-Radar Fusion for BEV Map and Object Segmentation
Jonas Schramm, Niclas Vödisch, Kürsat Petek, B Ravi Kiran, Senthil Yogamani, Wolfram Burgard, Abhinav Valada