Visual Navigation
Visual navigation research focuses on enabling robots and other agents to navigate environments using visual input, aiming to replicate the efficiency and adaptability of human spatial reasoning. Current efforts concentrate on developing robust and generalizable models, often employing deep learning architectures like transformers and convolutional neural networks, along with reinforcement learning techniques and novel map representations (e.g., Gaussian splatting, neural radiance fields). This field is significant for advancing robotics, assistive technologies for the visually impaired, and autonomous systems, with ongoing work addressing challenges like handling noisy sensor data, adapting to unseen environments, and improving efficiency and safety in complex scenarios.
Papers
Visual Navigation for Autonomous Vehicles: An Open-source Hands-on Robotics Course at MIT
Luca Carlone, Kasra Khosoussi, Vasileios Tzoumas, Golnaz Habibi, Markus Ryll, Rajat Talak, Jingnan Shi, Pasquale Antonante
SAMPLE-HD: Simultaneous Action and Motion Planning Learning Environment
Michal Nazarczuk, Tony Ng, Krystian Mikolajczyk