Explanation Quality
Explanation quality in artificial intelligence focuses on assessing the accuracy, usefulness, and trustworthiness of methods that explain AI model predictions. Current research emphasizes developing objective evaluation metrics beyond human judgments, often employing techniques like information bottleneck principles and comparing explanations against ground truth data or decision paths from simpler models (e.g., decision trees). This field is crucial for building trust in AI systems across various domains, particularly in high-stakes applications where understanding model decisions is paramount for safety, fairness, and accountability.
Papers
March 7, 2023
January 14, 2023
December 30, 2022
October 13, 2022
October 7, 2022
July 2, 2022
June 27, 2022
May 18, 2022
May 15, 2022
March 25, 2022
March 4, 2022