Interpretable Policy

Interpretable policy research focuses on developing machine learning models whose decision-making processes are transparent and understandable, addressing the "black box" problem of many AI systems. Current research emphasizes tree-based models, including decision trees and variations like Optimal MDP Decision Trees and Interpretable Continuous Control Trees, as well as neuro-symbolic approaches that combine neural networks with symbolic reasoning to create more explainable policies. This work is crucial for building trust in AI systems, particularly in high-stakes applications like autonomous driving and healthcare, where understanding the reasoning behind decisions is paramount for safety and accountability.

Papers