Better Explainability
Improving the explainability of machine learning models aims to make their decision-making processes more transparent and understandable, fostering trust and enabling better model debugging and refinement. Current research focuses on developing novel explanation methods, including those based on feature attribution, counterfactual examples, and structured argumentation, often applied to deep neural networks, transformers, and reinforcement learning agents. These advancements are crucial for deploying AI systems responsibly in high-stakes domains like healthcare and autonomous systems, where understanding model behavior is paramount.
Papers
November 7, 2024
October 17, 2024
September 4, 2024
July 21, 2024
June 26, 2024
June 20, 2024
June 7, 2024
May 20, 2024
January 29, 2024
January 25, 2024
October 27, 2023
October 24, 2023
October 19, 2023
October 3, 2023
July 29, 2023
July 19, 2023
June 27, 2023
June 5, 2023
May 26, 2023
April 12, 2023