Explanation Map

Explanation maps are visual representations designed to enhance the interpretability of complex deep learning models, particularly in computer vision. Current research focuses on developing methods to generate these maps for various architectures, including convolutional neural networks and vision transformers, often leveraging gradients, attention mechanisms, or iterative integration techniques to highlight the input features most influential in model predictions. These methods are evaluated using both reference-based metrics (comparing to human gaze patterns) and no-reference metrics (assessing stability and consistency), aiming to improve the trustworthiness and user understanding of AI systems. The ultimate goal is to bridge the "interpretation gap" between human understanding and the inner workings of these powerful, but often opaque, models.

Papers