Vision and Language Navigation
Vision-and-Language Navigation (VLN) focuses on enabling agents to navigate 3D environments by following natural language instructions, aiming to bridge the gap between visual perception and linguistic understanding. Current research emphasizes improving model efficiency (e.g., through knowledge distillation), exploring zero-shot navigation with large language models (LLMs) and incorporating safety mechanisms, and addressing challenges like instruction errors and robustness to environmental changes. This field is significant for advancing embodied AI and has potential applications in robotics, autonomous systems, and human-computer interaction.
Papers
June 9, 2022
April 20, 2022
April 6, 2022
March 29, 2022
March 28, 2022
March 22, 2022
March 10, 2022
March 5, 2022
February 23, 2022
February 14, 2022
January 26, 2022
November 28, 2021
November 14, 2021
November 10, 2021