Concept Explanation
Concept explanation research aims to make the predictions of complex machine learning models, particularly deep neural networks, more understandable and interpretable by humans. Current efforts focus on developing model-agnostic methods that identify and utilize high-level concepts (e.g., color, texture, object presence) to explain model decisions, often employing techniques like concept activation vectors and Bayesian estimation to improve robustness and accuracy. This work is crucial for building trust in AI systems, facilitating model debugging and improvement, and enabling effective human-computer collaboration in diverse applications such as plant disease classification and text analysis.
12papers
Papers
September 18, 2024
January 12, 2024
December 13, 2023
October 18, 2023
Concept-Guided Chain-of-Thought Prompting for Pairwise Comparison Scoring of Texts with Large Language Models
Patrick Y. Wu, Jonathan Nagler, Joshua A. Tucker, Solomon MessingFrom Neural Activations to Concepts: A Survey on Explaining Concepts in Neural Networks
Jae Hee Lee, Sergio Lanza, Stefan Wermter
September 15, 2023
November 19, 2022