Paper ID: 2206.02790

Improving Model Understanding and Trust with Counterfactual Explanations of Model Confidence

Thao Le, Tim Miller, Ronal Singh, Liz Sonenberg

In this paper, we show that counterfactual explanations of confidence scores help users better understand and better trust an AI model's prediction in human-subject studies. Showing confidence scores in human-agent interaction systems can help build trust between humans and AI systems. However, most existing research only used the confidence score as a form of communication, and we still lack ways to explain why the algorithm is confident. This paper also presents two methods for understanding model confidence using counterfactual explanation: (1) based on counterfactual examples; and (2) based on visualisation of the counterfactual space.

Submitted: Jun 6, 2022