Convincing Explanation
Convincing explanation in artificial intelligence focuses on generating human-understandable justifications for model outputs, aiming to improve trust and transparency in AI systems. Current research emphasizes the use of large language models (LLMs), often employing techniques like prompting and self-refinement to enhance the persuasiveness and faithfulness of generated explanations, while also investigating methods to detect and mitigate the risk of adversarially helpful, misleading explanations. This work is crucial for building reliable and trustworthy AI systems across various applications, from decision support systems to misinformation detection, by addressing the critical need for verifiable and understandable AI reasoning.
Papers
June 20, 2024
May 10, 2024
May 8, 2024
March 31, 2024
March 5, 2024
February 6, 2024
November 22, 2023
November 15, 2023
November 13, 2023
November 9, 2023
October 20, 2023
September 19, 2023
May 11, 2023
August 5, 2022
July 28, 2022