Saliency Method
Saliency methods aim to explain the decision-making processes of complex machine learning models, particularly deep neural networks, by identifying the input features most influential in generating a prediction. Current research focuses on developing more robust and reliable saliency algorithms, including those tailored for specific model architectures (e.g., transformers, convolutional neural networks) and data types (e.g., images, videos, time series, 3D data), as well as improving evaluation metrics for assessing the faithfulness of these explanations. The development of accurate and reliable saliency methods is crucial for enhancing the trustworthiness and interpretability of AI systems across various fields, from medical imaging to autonomous driving.
Papers
Signature Activation: A Sparse Signal View for Holistic Saliency
Jose Roberto Tello Ayala, Akl C. Fahed, Weiwei Pan, Eugene V. Pomerantsev, Patrick T. Ellinor, Anthony Philippakis, Finale Doshi-Velez
COSE: A Consistency-Sensitivity Metric for Saliency on Image Classification
Rangel Daroya, Aaron Sun, Subhransu Maji