Natural Language Explanation
Natural language explanation (NLE) research focuses on generating human-understandable explanations for AI model decisions, aiming to improve transparency, trust, and user understanding. Current efforts concentrate on developing methods to generate accurate, consistent, and faithful explanations using large language models (LLMs), often augmented with knowledge graphs or retrieval mechanisms, and evaluating these explanations using both automatic metrics and human assessments. This field is significant for enhancing the trustworthiness and usability of AI systems across diverse applications, from medicine and law to education and robotics, by bridging the gap between complex model outputs and human comprehension.
Papers
XForecast: Evaluating Natural Language Explanations for Time Series Forecasting
Taha Aksu, Chenghao Liu, Amrita Saha, Sarah Tan, Caiming Xiong, Doyen Sahoo
Towards Faithful Natural Language Explanations: A Study Using Activation Patching in Large Language Models
Wei Jie Yeo, Ranjan Satapathy, Erik Cambria
CasiMedicos-Arg: A Medical Question Answering Dataset Annotated with Explanatory Argumentative Structures
katerina Sviridova, Anar Yeginbergen, Ainara Estarrona, Elena Cabrio, Serena Villata, Rodrigo Agerri
LLaVA Needs More Knowledge: Retrieval Augmented Natural Language Generation with Knowledge Graph for Explaining Thoracic Pathologies
Ameer Hamza, Abdullah, Yong Hyun Ahn, Sungyoung Lee, Seong Tae Kim