Faithfulness Test
Faithfulness testing in explainable AI (XAI) aims to evaluate how accurately explanations of a model's predictions reflect its actual decision-making process. Current research focuses on developing robust and reliable metrics for assessing faithfulness across various model architectures, including large language models (LLMs) and convolutional neural networks (CNNs), often employing techniques like input perturbation, counterfactual generation, and adversarial attacks to probe explanation validity. These efforts are crucial for building trust in AI systems and ensuring their responsible deployment, particularly in high-stakes applications where understanding model behavior is paramount.
Papers
October 18, 2024
October 16, 2024
October 3, 2024
September 26, 2024
August 21, 2024
July 31, 2024
July 9, 2024
May 28, 2024
April 22, 2024
April 4, 2024
March 29, 2024
November 13, 2023
November 3, 2023
November 2, 2023
October 18, 2023
October 11, 2023
August 17, 2023
August 13, 2023