Imitation Policy
Imitation learning aims to train agents to mimic expert behavior from observational data, bypassing the need for explicit reward functions. Current research emphasizes improving robustness and generalization of learned policies, focusing on techniques like offline-to-online finetuning, selective imitation from large datasets, and adversarial methods to address compounding errors and gradient explosion issues in algorithms such as behavior cloning and Generative Adversarial Imitation Learning (GAIL). These advancements are crucial for deploying reliable imitation policies in safety-critical applications like robotics and autonomous systems, where generalization to unseen scenarios is paramount.
Papers
Memory-Consistent Neural Networks for Imitation Learning
Kaustubh Sridhar, Souradeep Dutta, Dinesh Jayaraman, James Weimer, Insup Lee
Imitator Learning: Achieve Out-of-the-Box Imitation Ability in Variable Environments
Xiong-Hui Chen, Junyin Ye, Hang Zhao, Yi-Chen Li, Haoran Shi, Yu-Yan Xu, Zhihao Ye, Si-Hang Yang, Anqi Huang, Kai Xu, Zongzhang Zhang, Yang Yu