Interpretable Concept
Interpretable concept research aims to make the decision-making processes of complex machine learning models, particularly deep learning models, more transparent and understandable. Current efforts focus on developing model architectures, such as concept bottleneck models and self-explaining neural networks, that incorporate human-interpretable concepts into their design or post-hoc explanation methods that leverage vision-language models or unsupervised concept discovery. This work is crucial for building trust in AI systems across various applications, from medical diagnosis to autonomous driving, by providing insights into model behavior and enabling more reliable and verifiable predictions.
Papers
October 28, 2024
October 21, 2024
October 6, 2024
August 16, 2024
August 5, 2024
July 22, 2024
July 19, 2024
July 9, 2024
July 4, 2024
June 28, 2024
June 27, 2024
June 12, 2024
May 24, 2024
April 18, 2024
April 2, 2024
February 14, 2024
January 16, 2024
January 12, 2024
December 21, 2023