Black Box Model
Black box models, characterized by their opaque internal workings, pose challenges in understanding their decision-making processes, hindering trust and accountability. Current research focuses on improving interpretability through methods like generalized additive models (GAMs) and surrogate models, as well as addressing vulnerabilities to adversarial attacks and biases through techniques such as explanation-driven attacks and robust defense mechanisms. This work is crucial for building trust in AI systems across various applications, from medical diagnosis to autonomous driving, by enhancing transparency and mitigating potential risks associated with unpredictable model behavior.
Papers
Only My Model On My Data: A Privacy Preserving Approach Protecting one Model and Deceiving Unauthorized Black-Box Models
Weiheng Chai, Brian Testa, Huantao Ren, Asif Salekin, Senem Velipasalar
Implementing local-explainability in Gradient Boosting Trees: Feature Contribution
Ángel Delgado-Panadero, Beatriz Hernández-Lorca, María Teresa García-Ordás, José Alberto Benítez-Andrades
The Mirrored Influence Hypothesis: Efficient Data Influence Estimation by Harnessing Forward Passes
Myeongseob Ko, Feiyang Kang, Weiyan Shi, Ming Jin, Zhou Yu, Ruoxi Jia
Explainable AI for survival analysis: a median-SHAP approach
Lucile Ter-Minassian, Sahra Ghalebikesabi, Karla Diaz-Ordaz, Chris Holmes
Explainable data-driven modeling via mixture of experts: towards effective blending of grey and black-box models
Jessica Leoni, Valentina Breschi, Simone Formentin, Mara Tanelli