Black Box Model

Black box models, characterized by their opaque internal workings, pose challenges in understanding their decision-making processes, hindering trust and accountability. Current research focuses on improving interpretability through methods like generalized additive models (GAMs) and surrogate models, as well as addressing vulnerabilities to adversarial attacks and biases through techniques such as explanation-driven attacks and robust defense mechanisms. This work is crucial for building trust in AI systems across various applications, from medical diagnosis to autonomous driving, by enhancing transparency and mitigating potential risks associated with unpredictable model behavior.

Papers