Counterfactual Explanation
Counterfactual explanations (CFEs) aim to enhance the interpretability of machine learning models by showing how minimal input changes would alter predictions. Current research focuses on developing robust and efficient CFE generation methods across various model types, including deep learning architectures like variational autoencoders and diffusion models, and for diverse data modalities such as images, time series, and text. This work is significant because CFEs improve model transparency and trustworthiness, fostering greater user understanding and facilitating the responsible deployment of AI in high-stakes applications like healthcare and finance.
Papers
Overlap Number of Balls Model-Agnostic CounterFactuals (ONB-MACF): A Data-Morphology-based Counterfactual Generation Method for Trustworthy Artificial Intelligence
José Daniel Pascual-Triana, Alberto Fernández, Javier Del Ser, Francisco Herrera
Exploring Commonalities in Explanation Frameworks: A Multi-Domain Survey Analysis
Eduard Barbu, Marharytha Domnich, Raul Vicente, Nikos Sakkas, André Morim
Model Reconstruction Using Counterfactual Explanations: A Perspective From Polytope Theory
Pasan Dissanayake, Sanghamitra Dutta
Relevant Irrelevance: Generating Alterfactual Explanations for Image Classifiers
Silvan Mertes, Tobias Huber, Christina Karle, Katharina Weitz, Ruben Schlagowski, Cristina Conati, Elisabeth André