Inverse Q Learning

Inverse Q-learning aims to infer reward functions from observed expert behavior, essentially reversing the standard reinforcement learning process. Current research focuses on developing robust algorithms, such as hierarchical inverse Q-learning (HIQL) and least squares inverse Q-learning (LS-IQ), that address challenges like handling absorbing states and learning from diverse, potentially unlabeled data. These advancements are improving the interpretability of learned behaviors, particularly in applications like understanding animal decision-making and enabling more effective offline reinforcement learning for complex tasks, such as goal-conditioned robotics. The resulting insights are valuable for both scientific understanding of decision-making processes and the development of more efficient and robust AI systems.

Papers