Multimodal Misinformation
Multimodal misinformation, the spread of false information across multiple media formats (text, image, video, audio), is a growing concern demanding robust detection methods. Current research focuses on developing automated fact-checking systems using large vision-language models (LVLMs) and large language models (LLMs), often incorporating techniques like contrastive learning, knowledge distillation, and external knowledge augmentation to improve accuracy and interpretability. These advancements aim to address the challenges posed by increasingly sophisticated misinformation campaigns and contribute to the development of more resilient and trustworthy information ecosystems.
Papers
Multimodal Automated Fact-Checking: A Survey
Mubashara Akhtar, Michael Schlichtkrull, Zhijiang Guo, Oana Cocarascu, Elena Simperl, Andreas Vlachos
FACTIFY3M: A Benchmark for Multimodal Fact Verification with Explainability through 5W Question-Answering
Megha Chakraborty, Khushbu Pahwa, Anku Rani, Shreyas Chatterjee, Dwip Dalal, Harshit Dave, Ritvik G, Preethi Gurumurthy, Adarsh Mahor, Samahriti Mukherjee, Aditya Pakala, Ishan Paul, Janvita Reddy, Arghya Sarkar, Kinjal Sensharma, Aman Chadha, Amit P. Sheth, Amitava Das