Multimodal Chart
Multimodal chart research focuses on developing computational methods to understand and interact with charts containing diverse data types like text, images, and numerical data. Current efforts concentrate on leveraging large language models (LLMs) and vision transformers (ViTs) for tasks such as question answering, sentiment analysis, and information extraction from charts, often employing techniques like contrastive learning and knowledge distillation to improve model performance. This field is significant for advancing human-computer interaction and enabling more sophisticated analysis of complex visual data across various domains, including healthcare, scientific literature, and social media.
Papers
December 31, 2023
December 18, 2023
November 24, 2023
November 15, 2023
September 7, 2023
August 29, 2023
April 19, 2023
March 31, 2023
March 6, 2023
February 1, 2023
December 15, 2022
November 15, 2022
November 11, 2022
August 16, 2022
May 9, 2022
March 23, 2022