Plausible Explanation
Plausible explanation in AI focuses on generating understandable and trustworthy justifications for model predictions, aiming to bridge the gap between complex algorithms and human comprehension. Current research emphasizes diverse explanation methods, including counterfactual examples, feature attributions (often aggregated using optimization techniques), and natural language explanations generated via large language models (LLMs) or other architectures like variational autoencoders. This field is crucial for building trust in AI systems, particularly in high-stakes domains like medicine, and for improving model transparency and accountability by identifying and mitigating biases or spurious correlations.
Papers
March 30, 2023
March 8, 2023
December 22, 2022
December 18, 2022
November 25, 2022
November 14, 2022
October 12, 2022
September 2, 2022
August 12, 2022
August 5, 2022
July 17, 2022
July 4, 2022
June 15, 2022
June 9, 2022
May 11, 2022