Attribution Quality
Attribution quality in machine learning focuses on evaluating how well model explanations, such as attribution maps, reflect the true decision-making process of the model. Current research emphasizes developing robust evaluation protocols that address biases and inconsistencies in existing methods, exploring the impact of model architecture on explanation quality, and investigating the use of planning and calibration techniques to improve attribution accuracy across various model types, including vision, language, and generative models. Improved attribution quality is crucial for building trust in AI systems, facilitating model debugging and refinement, and enabling more reliable and interpretable AI applications.
Papers
October 16, 2024
July 16, 2024
April 4, 2024
January 19, 2024
November 6, 2022
October 10, 2022
February 17, 2022