Explainability Tool
Explainability tools aim to make the decision-making processes of complex machine learning models, particularly "black box" models like deep neural networks and large language models, more transparent and understandable. Current research focuses on developing methods to provide both local (instance-specific) and global explanations, employing techniques like counterfactual analysis, feature attribution, and interactive dialogue systems. These tools are crucial for building trust in AI systems, improving model debugging and fairness, and enabling responsible deployment across diverse applications, including healthcare, finance, and security.
Papers
August 22, 2024
August 21, 2024
April 3, 2024
February 16, 2024
January 23, 2024
November 24, 2023
November 22, 2023
November 20, 2023
September 25, 2023
May 26, 2023
April 26, 2023
September 8, 2022
July 8, 2022
April 14, 2022
December 9, 2021