Interpretable Concept
Interpretable concept research aims to make the decision-making processes of complex machine learning models, particularly deep learning models, more transparent and understandable. Current efforts focus on developing model architectures, such as concept bottleneck models and self-explaining neural networks, that incorporate human-interpretable concepts into their design or post-hoc explanation methods that leverage vision-language models or unsupervised concept discovery. This work is crucial for building trust in AI systems across various applications, from medical diagnosis to autonomous driving, by providing insights into model behavior and enabling more reliable and verifiable predictions.
Papers
November 15, 2023
October 25, 2023
October 23, 2023
September 14, 2023
August 24, 2023
July 15, 2023
May 27, 2023
May 1, 2023
April 12, 2023
April 10, 2023
March 22, 2023
February 9, 2023
December 14, 2022
November 19, 2022
August 17, 2022
August 4, 2022
June 28, 2022
April 1, 2022
March 25, 2022