Better Explainability
Improving the explainability of machine learning models aims to make their decision-making processes more transparent and understandable, fostering trust and enabling better model debugging and refinement. Current research focuses on developing novel explanation methods, including those based on feature attribution, counterfactual examples, and structured argumentation, often applied to deep neural networks, transformers, and reinforcement learning agents. These advancements are crucial for deploying AI systems responsibly in high-stakes domains like healthcare and autonomous systems, where understanding model behavior is paramount.
Papers
March 17, 2023
December 2, 2022
November 22, 2022
October 13, 2022
June 26, 2022
June 15, 2022
May 25, 2022
April 14, 2022
February 15, 2022
December 29, 2021
December 7, 2021