Natural Language Explanation
Natural language explanation (NLE) research focuses on generating human-understandable explanations for AI model decisions, aiming to improve transparency, trust, and user understanding. Current efforts concentrate on developing methods to generate accurate, consistent, and faithful explanations using large language models (LLMs), often augmented with knowledge graphs or retrieval mechanisms, and evaluating these explanations using both automatic metrics and human assessments. This field is significant for enhancing the trustworthiness and usability of AI systems across diverse applications, from medicine and law to education and robotics, by bridging the gap between complex model outputs and human comprehension.
Papers
Enhancing adversarial robustness in Natural Language Inference using explanations
Alexandros Koulakos, Maria Lymperaiou, Giorgos Filandrianos, Giorgos Stamou
Cross-Refine: Improving Natural Language Explanation Generation by Learning in Tandem
Qianli Wang, Tatiana Anikina, Nils Feldhus, Simon Ostermann, Sebastian Möller, Vera Schmitt
Explainable machine learning multi-label classification of Spanish legal judgements
Francisco de Arriba-Pérez, Silvia García-Méndez, Francisco J. González-Castaño, Jaime González-González
How Ready Are Generative Pre-trained Large Language Models for Explaining Bengali Grammatical Errors?
Subhankar Maity, Aniket Deroy, Sudeshna Sarkar