Hierarchical Attribution
Hierarchical attribution aims to explain complex model predictions by decomposing them into contributions from different levels of features or sub-populations, providing insights into model behavior and identifying biases. Current research focuses on developing efficient algorithms for generating these hierarchical explanations, including methods leveraging hyperbolic spaces and attention mechanisms to capture feature interactions and improve explainability in diverse applications like natural language processing and medical image segmentation. This work is crucial for enhancing the transparency and trustworthiness of machine learning models, particularly in high-stakes domains where understanding model decisions is paramount.
Papers
October 13, 2024
July 25, 2024
July 23, 2024
March 25, 2024
March 15, 2024
October 24, 2022
October 4, 2022