Hierarchical Attribution

Hierarchical attribution aims to explain complex model predictions by decomposing them into contributions from different levels of features or sub-populations, providing insights into model behavior and identifying biases. Current research focuses on developing efficient algorithms for generating these hierarchical explanations, including methods leveraging hyperbolic spaces and attention mechanisms to capture feature interactions and improve explainability in diverse applications like natural language processing and medical image segmentation. This work is crucial for enhancing the transparency and trustworthiness of machine learning models, particularly in high-stakes domains where understanding model decisions is paramount.

Papers