Visual Question Generation
Visual Question Generation (VQG) focuses on automatically creating natural language questions from images, aiming to mimic human questioning behavior and improve human-computer interaction. Current research emphasizes generating more diverse and relevant questions by incorporating contextual information like answers, regions of interest within the image, and external knowledge bases, often leveraging transformer-based encoder-decoder architectures and contrastive learning methods. This field is significant for advancing multimodal AI, enabling more sophisticated question answering systems, and facilitating applications in education, conversational agents, and data cleansing through improved data annotation and analysis.
Papers
July 6, 2024
February 20, 2024
January 19, 2024
January 18, 2024
October 23, 2023
October 12, 2023
September 28, 2023
June 11, 2023
May 28, 2023
November 23, 2022
November 14, 2022
October 13, 2022
August 3, 2022
May 19, 2022
March 15, 2022