Inverse Reinforcement Learning
Inverse reinforcement learning (IRL) aims to infer an agent's reward function from observations of its behavior, essentially reverse-engineering the decision-making process. Current research emphasizes improving the robustness and efficiency of IRL algorithms, particularly in handling noisy or incomplete data, diverse expert policies, and non-Markovian rewards, often employing techniques like maximum entropy IRL, Bayesian IRL, and various model-predictive control methods. These advancements are crucial for applications such as robotics, autonomous driving, and human-computer interaction, where learning from human demonstrations or preferences is essential for safe and effective system design. Furthermore, research is actively addressing challenges like scalability to large state spaces and the transferability of learned reward functions to new environments.
Papers
Pruning the Path to Optimal Care: Identifying Systematically Suboptimal Medical Decision-Making with Inverse Reinforcement Learning
Inko Bovenzi, Adi Carmel, Michael Hu, Rebecca M. Hurwitz, Fiona McBride, Leo Benac, José Roberto Tello Ayala, Finale Doshi-Velez
Inverse Transition Learning: Learning Dynamics from Demonstrations
Leo Benac, Abhishek Sharma, Sonali Parbhoo, Finale Doshi-Velez
UNIQ: Offline Inverse Q-learning for Avoiding Undesirable Demonstrations
Huy Hoang, Tien Mai, Pradeep Varakantham
Rethinking Adversarial Inverse Reinforcement Learning: From the Angles of Policy Imitation and Transferable Reward Recovery
Yangchun Zhang, Wang Zhou, Yirui Zhou
Offline Inverse Constrained Reinforcement Learning for Safe-Critical Decision Making in Healthcare
Nan Fang, Guiliang Liu, Wei Gong