Counterfactual Explanation
Counterfactual explanations (CFEs) aim to enhance the interpretability of machine learning models by showing how minimal input changes would alter predictions. Current research focuses on developing robust and efficient CFE generation methods across various model types, including deep learning architectures like variational autoencoders and diffusion models, and for diverse data modalities such as images, time series, and text. This work is significant because CFEs improve model transparency and trustworthiness, fostering greater user understanding and facilitating the responsible deployment of AI in high-stakes applications like healthcare and finance.
Papers
Motif-guided Time Series Counterfactual Explanations
Peiyu Li, Soukaina Filali Boubrahimi, Shah Muhammad Hamdi
Generating counterfactual explanations of tumor spatial proteomes to discover effective strategies for enhancing immune infiltration
Zitong Jerry Wang, Alexander M. Xu, Aman Bhargava, Matt W. Thomson
Augmentation by Counterfactual Explanation -- Fixing an Overconfident Classifier
Sumedha Singla, Nihal Murali, Forough Arabshahi, Sofia Triantafyllou, Kayhan Batmanghelich
A Survey on Graph Counterfactual Explanations: Definitions, Methods, Evaluation, and Research Challenges
Mario Alfonso Prado-Romero, Bardh Prenkaj, Giovanni Stilo, Fosca Giannotti
The privacy issue of counterfactual explanations: explanation linkage attacks
Sofie Goethals, Kenneth Sörensen, David Martens
Redefining Counterfactual Explanations for Reinforcement Learning: Overview, Challenges and Opportunities
Jasmina Gajcin, Ivana Dusparic
Diffusion Visual Counterfactual Explanations
Maximilian Augustin, Valentyn Boreiko, Francesco Croce, Matthias Hein