Model Interpretation
Model interpretation focuses on understanding how machine learning models arrive at their predictions, aiming to increase transparency and trust in their use. Current research emphasizes distinguishing between model-centric ("sensitive") and task-centric ("decisive") patterns in interpretations, exploring various methods including post-hoc and self-interpretable approaches applied to diverse architectures like convolutional neural networks and tree ensembles. This field is crucial for ensuring responsible AI deployment across scientific domains and practical applications, particularly in high-stakes areas like healthcare and finance, by providing insights into model behavior and identifying potential biases or flaws.
Papers
October 21, 2024
October 2, 2024
October 1, 2024
June 30, 2024
May 31, 2024
April 3, 2024
January 27, 2024
November 27, 2023
October 25, 2023
May 17, 2023
February 8, 2023
February 7, 2023
September 12, 2022
September 1, 2022
June 12, 2022
June 11, 2022