Plausible Explanation
Plausible explanation in AI focuses on generating understandable and trustworthy justifications for model predictions, aiming to bridge the gap between complex algorithms and human comprehension. Current research emphasizes diverse explanation methods, including counterfactual examples, feature attributions (often aggregated using optimization techniques), and natural language explanations generated via large language models (LLMs) or other architectures like variational autoencoders. This field is crucial for building trust in AI systems, particularly in high-stakes domains like medicine, and for improving model transparency and accountability by identifying and mitigating biases or spurious correlations.
Papers
October 18, 2024
July 30, 2024
July 22, 2024
July 12, 2024
June 10, 2024
June 7, 2024
June 5, 2024
May 28, 2024
May 11, 2024
March 8, 2024
February 7, 2024
February 5, 2024
February 1, 2024
January 25, 2024
December 17, 2023
November 29, 2023
September 25, 2023
September 19, 2023
June 24, 2023