Black Box Model
Black box models, characterized by their opaque internal workings, pose challenges in understanding their decision-making processes, hindering trust and accountability. Current research focuses on improving interpretability through methods like generalized additive models (GAMs) and surrogate models, as well as addressing vulnerabilities to adversarial attacks and biases through techniques such as explanation-driven attacks and robust defense mechanisms. This work is crucial for building trust in AI systems across various applications, from medical diagnosis to autonomous driving, by enhancing transparency and mitigating potential risks associated with unpredictable model behavior.
Papers
Uncertainty Guarantees on Automated Precision Weeding using Conformal Prediction
Paul Melki (IMS), Lionel Bombrun (IMS), Boubacar Diallo, Jérôme Dias, Jean-Pierre da Costa (IMS)
Neural Probabilistic Circuits: Enabling Compositional and Interpretable Predictions through Logical Reasoning
Weixin Chen, Simon Yu, Huajie Shao, Lui Sha, Han Zhao
BAMBA: A Bimodal Adversarial Multi-Round Black-Box Jailbreak Attacker for LVLMs
Ruoxi Cheng, Yizhong Ding, Shuirong Cao, Shaowei Yuan, Zhiqiang Wang, Xiaojun Jia
DREAM: Domain-agnostic Reverse Engineering Attributes of Black-box Model
Rongqing Li, Jiaqi Yu, Changsheng Li, Wenhan Luo, Ye Yuan, Guoren Wang