Explanation Method
Explanation methods aim to make the decision-making processes of complex machine learning models more transparent and understandable. Current research focuses on improving the faithfulness, stability, and user-friendliness of explanations, exploring various approaches including SHAP, LIME, gradient-based methods, and the use of large language models to generate more natural and engaging explanations. This work is crucial for building trust in AI systems, particularly in high-stakes applications like healthcare and finance, and for facilitating better model debugging and design. A key challenge remains developing robust evaluation metrics that capture the multifaceted nature of explanation quality and its impact on human understanding.
Papers
SoK: Modeling Explainability in Security Analytics for Interpretability, Trustworthiness, and Usability
Dipkamal Bhusal, Rosalyn Shin, Ajay Ashok Shewale, Monish Kumar Manikya Veerabhadran, Michael Clifford, Sara Rampazzi, Nidhi Rastogi
Computing Rule-Based Explanations by Leveraging Counterfactuals
Zixuan Geng, Maximilian Schleich, Dan Suciu