Explainability Method
Explainability methods aim to make the decision-making processes of complex machine learning models, particularly deep neural networks and large language models, more transparent and understandable. Current research focuses on developing and evaluating methods that assess the faithfulness and plausibility of explanations, often using techniques like counterfactual generation, attribution methods (e.g., SHAP, LIME, Grad-CAM), and concept-based approaches. This work is crucial for building trust in AI systems across diverse applications, from medical diagnosis to autonomous vehicles, by providing insights into model behavior and identifying potential biases.
Papers
June 2, 2023
June 1, 2023
May 25, 2023
April 28, 2023
April 18, 2023
March 15, 2023
March 6, 2023
February 23, 2023
November 22, 2022
November 18, 2022
November 9, 2022
November 3, 2022
October 13, 2022
October 11, 2022
August 25, 2022
August 12, 2022
August 5, 2022
July 5, 2022