Human Interpretability
Human interpretability in artificial intelligence focuses on making AI models' decision-making processes understandable to humans, thereby increasing trust and facilitating effective use in sensitive applications. Current research emphasizes developing inherently interpretable models, such as those incorporating biological knowledge or leveraging simpler architectures like linear classifiers and kernel methods, alongside post-hoc explanation techniques for existing "black box" models. This work is crucial for advancing AI's adoption in fields like medicine and healthcare, where understanding the reasoning behind predictions is paramount for responsible and effective implementation.
Papers
October 21, 2024
July 5, 2024
June 24, 2023
May 24, 2023
October 19, 2022
October 3, 2022
July 2, 2022
June 27, 2022
January 18, 2022
December 6, 2021