Model Interpretability
Model interpretability aims to make the decision-making processes of complex machine learning models transparent and understandable. Current research focuses on developing both inherently interpretable models, such as generalized additive models and rule-based systems, and post-hoc methods that explain the predictions of black-box models, often using techniques like SHAP values, Grad-CAM, and various attention mechanisms applied to architectures like transformers and neural networks. This field is crucial for building trust in AI systems, particularly in high-stakes domains like healthcare and finance, and for facilitating the responsible development and deployment of machine learning technologies.
Papers
December 15, 2022
December 14, 2022
October 24, 2022
October 17, 2022
September 30, 2022
September 29, 2022
September 21, 2022
September 7, 2022
August 1, 2022
June 12, 2022
May 25, 2022
May 3, 2022
April 19, 2022
April 12, 2022
March 28, 2022
February 14, 2022
January 24, 2022
November 16, 2021