Interpretable Model
Interpretable models aim to create machine learning systems whose decision-making processes are transparent and understandable to humans, addressing the "black box" problem of many high-performing models. Current research focuses on developing inherently interpretable architectures like generalized additive models (GAMs), decision trees, rule lists, and symbolic regression, as well as post-hoc explanation methods for existing models, such as SHAP and LIME. This emphasis on interpretability is driven by the need for trust, accountability, and the ability to gain insights from complex data in fields ranging from healthcare and finance to scientific discovery, where understanding model decisions is crucial for effective application and responsible use. The development of more accurate and efficient methods for creating and evaluating interpretable models is a major focus of ongoing research.
Papers
Banach-Tarski Embeddings and Transformers
Joshua Maher
Cross-domain feature disentanglement for interpretable modeling of tumor microenvironment impact on drug response
Jia Zhai, Hui Liu
Interpretable by Design: Wrapper Boxes Combine Neural Performance with Faithful Explanations
Yiheng Su, Juni Jessy Li, Matthew Lease
Extreme sparsification of physics-augmented neural networks for interpretable model discovery in mechanics
Jan N. Fuhg, Reese E. Jones, Nikolaos Bouklas
The Blame Problem in Evaluating Local Explanations, and How to Tackle it
Amir Hossein Akhavan Rahnama
A Quantitatively Interpretable Model for Alzheimer's Disease Prediction Using Deep Counterfactuals
Kwanseok Oh, Da-Woon Heo, Ahmad Wisnu Mulyadi, Wonsik Jung, Eunsong Kang, Kun Ho Lee, Heung-Il Suk