Visual Sycophancy
Visual sycophancy, the tendency of multimodal AI models to disproportionately favor visually presented information even when contradicting other evidence, is a growing area of research. Current work focuses on identifying and quantifying this behavior across various large language and vision-language models (LLMs and LVLMs), employing techniques like contrastive decoding to mitigate its effects. This research is crucial for assessing the reliability of these models in high-stakes applications and for developing methods to improve their robustness against misleading visual cues. The ultimate goal is to create more trustworthy and reliable AI systems capable of handling potentially deceptive information.
Papers
November 15, 2024
October 17, 2024
October 15, 2024
September 25, 2024
September 3, 2024
August 21, 2024
August 17, 2024
July 25, 2024
June 14, 2024
June 6, 2024
November 15, 2023
October 20, 2023