Visual Context
Visual context research explores how incorporating visual information improves the performance of AI models in various tasks, primarily aiming to enhance understanding and reasoning capabilities beyond simple image recognition. Current research focuses on developing multimodal models that integrate visual and textual data, often employing transformer architectures and large language models (LLMs) to process complex visual scenes and generate contextually relevant outputs. This field is significant because it addresses limitations in current AI systems, leading to improvements in applications such as image captioning, visual question answering, and autonomous driving, where understanding the visual environment is crucial.
Papers
November 12, 2024
October 23, 2024
October 15, 2024
October 5, 2024
August 13, 2024
August 12, 2024
July 17, 2024
July 6, 2024
June 24, 2024
June 17, 2024
May 21, 2024
May 8, 2024
April 18, 2024
March 26, 2024
March 5, 2024
February 28, 2024
February 22, 2024
February 13, 2024