Feature Importance
Feature importance quantifies the contribution of individual input features to a machine learning model's predictions, aiming to enhance model interpretability and facilitate better decision-making. Current research focuses on developing robust and reliable methods, particularly addressing challenges posed by complex models like deep neural networks and exploring techniques like Shapley values, permutation feature importance, and various game-theoretic approaches. This work is crucial for improving trust in AI systems across diverse fields, from healthcare and finance to autonomous vehicles, by providing insights into model behavior and enabling more informed feature engineering and selection. Furthermore, research is actively investigating the relationship between feature importance and model performance, as well as the development of fair and unbiased feature importance measures.
Papers
Disentangling Interactions and Dependencies in Feature Attribution
Gunnar König, Eric Günther, Ulrike von Luxburg
EXACFS -- A CIL Method to mitigate Catastrophic Forgetting
S Balasubramanian, M Sai Subramaniam, Sai Sriram Talasu, P Yedu Krishna, Manepalli Pranav Phanindra Sai, Ravi Mukkamala, Darshan Gera