Local Attribution
Local attribution methods aim to explain individual predictions made by complex machine learning models, identifying which input features or training data points most influenced the outcome. Current research focuses on improving the efficiency and accuracy of these methods, particularly for challenging scenarios like low signal-to-noise data and diverse model architectures (including neural networks and Siamese encoders), often employing techniques like surrogate modeling and integrated gradients. This work is crucial for enhancing the transparency and trustworthiness of machine learning systems across various applications, from image generation to time series analysis, by providing insights into model behavior and potential biases.
Papers
June 17, 2024
February 19, 2024
October 9, 2023
July 14, 2023
June 15, 2023
September 6, 2022
May 31, 2022
May 11, 2022