High Explainability
High explainability in artificial intelligence (AI) aims to make the decision-making processes of complex models, such as large language models and deep neural networks, more transparent and understandable. Current research focuses on developing both intrinsic (built-in) and post-hoc (added after training) explainability methods, often employing techniques like attention mechanisms, feature attribution, and counterfactual examples to interpret model outputs across various modalities (text, images, audio). This pursuit is crucial for building trust in AI systems, particularly in high-stakes domains like medicine and finance, and for ensuring fairness, accountability, and responsible AI development.
Papers
ADVISE: ADaptive Feature Relevance and VISual Explanations for Convolutional Neural Networks
Mohammad Mahdi Dehshibi, Mona Ashtari-Majlan, Gereziher Adhane, David Masip
Satellite Image and Machine Learning based Knowledge Extraction in the Poverty and Welfare Domain
Ola Hall, Mattias Ohlsson, Thortseinn Rögnvaldsson
Quantus: An Explainable AI Toolkit for Responsible Evaluation of Neural Network Explanations and Beyond
Anna Hedström, Leander Weber, Dilyara Bareeva, Daniel Krakowczyk, Franz Motzkus, Wojciech Samek, Sebastian Lapuschkin, Marina M. -C. Höhne
Design of Explainability Module with Experts in the Loop for Visualization and Dynamic Adjustment of Continual Learning
Yujiang He, Zhixin Huang, Bernhard Sick
Interpretable pipelines with evolutionarily optimized modules for RL tasks with visual inputs
Leonardo Lucio Custode, Giovanni Iacca
Investigating Explainability of Generative AI for Code through Scenario-based Design
Jiao Sun, Q. Vera Liao, Michael Muller, Mayank Agarwal, Stephanie Houde, Kartik Talamadupula, Justin D. Weisz