Black Box Model
Black box models, characterized by their opaque internal workings, pose challenges in understanding their decision-making processes, hindering trust and accountability. Current research focuses on improving interpretability through methods like generalized additive models (GAMs) and surrogate models, as well as addressing vulnerabilities to adversarial attacks and biases through techniques such as explanation-driven attacks and robust defense mechanisms. This work is crucial for building trust in AI systems across various applications, from medical diagnosis to autonomous driving, by enhancing transparency and mitigating potential risks associated with unpredictable model behavior.
Papers
Explainable AI for survival analysis: a median-SHAP approach
Lucile Ter-Minassian, Sahra Ghalebikesabi, Karla Diaz-Ordaz, Chris Holmes
Explainable data-driven modeling via mixture of experts: towards effective blending of grey and black-box models
Jessica Leoni, Valentina Breschi, Simone Formentin, Mara Tanelli