Interpretable Algorithm
Interpretable algorithms aim to make the decision-making processes of machine learning models transparent and understandable, enhancing trust and accountability. Current research focuses on developing methods to reverse-engineer existing models (e.g., transformers) to reveal their underlying algorithms, as well as designing inherently interpretable models from the outset, often employing techniques like concept bottlenecks and incorporating prior knowledge through optimization frameworks. This field is crucial for building trust in AI systems across various domains, particularly in high-stakes applications like healthcare, where understanding model decisions is paramount for responsible deployment.
Papers
August 18, 2024
June 17, 2024
June 14, 2024
May 17, 2024
April 22, 2024
February 26, 2024
December 5, 2023
November 30, 2023
October 12, 2023
September 7, 2023
July 6, 2023
May 15, 2023
February 6, 2023
May 31, 2022
May 20, 2022
May 11, 2022