Model Hallucination
Model hallucination, the generation of factually incorrect or nonsensical outputs by large language models (LLMs) and other AI systems, is a significant challenge hindering their reliable deployment. Current research focuses on developing methods to detect and mitigate hallucinations, employing techniques like retrieval-augmented generation (RAG), contrastive decoding, and targeted instruction tuning across various model architectures, including LLMs and large vision-language models (LVLMs). These efforts aim to improve the accuracy and trustworthiness of AI systems, impacting diverse applications from medical report generation to virtual try-on technologies. The development of robust hallucination detection and mitigation strategies is crucial for building reliable and responsible AI systems.
Papers
CoMT: Chain-of-Medical-Thought Reduces Hallucination in Medical Report Generation
Yue Jiang, Jiawei Chen, Dingkang Yang, Mingcheng Li, Shunli Wang, Tong Wu, Ke Li, Lihua Zhang
Small Agent Can Also Rock! Empowering Small Language Models as Hallucination Detector
Xiaoxue Cheng, Junyi Li, Wayne Xin Zhao, Hongzhi Zhang, Fuzheng Zhang, Di Zhang, Kun Gai, Ji-Rong Wen
TimeChara: Evaluating Point-in-Time Character Hallucination of Role-Playing Large Language Models
Jaewoo Ahn, Taehyun Lee, Junyoung Lim, Jin-Hwa Kim, Sangdoo Yun, Hwaran Lee, Gunhee Kim
RITUAL: Random Image Transformations as a Universal Anti-hallucination Lever in LVLMs
Sangmin Woo, Jaehyuk Jang, Donguk Kim, Yubin Choi, Changick Kim
IBD: Alleviating Hallucinations in Large Vision-Language Models via Image-Biased Decoding
Lanyun Zhu, Deyi Ji, Tianrun Chen, Peng Xu, Jieping Ye, Jun Liu
Collaborative decoding of critical tokens for boosting factuality of large language models
Lifeng Jin, Baolin Peng, Linfeng Song, Haitao Mi, Ye Tian, Dong Yu