Concept Bottleneck Model
Concept Bottleneck Models (CBMs) aim to enhance the interpretability of deep learning models by incorporating human-understandable concepts into the prediction process, thereby bridging the gap between complex model outputs and human comprehension. Current research focuses on improving CBM accuracy, addressing security vulnerabilities like backdoor attacks, and developing methods for automated concept discovery and selection, often leveraging vision-language models like CLIP. This work is significant because it strives to create more trustworthy and reliable AI systems, particularly in high-stakes domains like medicine, where understanding model decisions is crucial for both performance and user acceptance.
Papers
Guarding the Gate: ConceptGuard Battles Concept-Level Backdoors in Concept Bottleneck Models
Songning Lai, Yu Huang, Jiayu Yang, Gaoxiang Huang, Wenshuo Chen, Yutao Yue
Learning New Concepts, Remembering the Old: A Novel Continual Learning
Songning Lai, Mingqian Liao, Zhangyi Hu, Jiayu Yang, Wenshuo Chen, Yutao Yue
Tree-Based Leakage Inspection and Control in Concept Bottleneck Models
Angelos Ragkousis, Sonali Parbhoo
Scalable Mechanistic Neural Networks
Jiale Chen, Dingling Yao, Adeel Pervez, Dan Alistarh, Francesco Locatello
Enforcing Interpretability in Time Series Transformers: A Concept Bottleneck Framework
Angela van Sprang, Erman Acar, Willem Zuidema