Imitation Learning
Imitation learning aims to train agents to mimic expert behavior by learning from observational data, primarily focusing on efficiently transferring complex skills from humans or other advanced controllers to robots. Current research emphasizes improving data efficiency through techniques like active learning, data augmentation, and leveraging large language models to provide richer context and handle failures. This field is crucial for advancing robotics, autonomous driving, and other areas requiring complex control policies, as it offers a more data-driven and potentially less labor-intensive approach than traditional programming methods.
Papers
CGD: Constraint-Guided Diffusion Policies for UAV Trajectory Planning
Kota Kondo, Andrea Tagliabue, Xiaoyi Cai, Claudius Tewari, Olivia Garcia, Marcos Espitia-Alvarez, Jonathan P. How
IntervenGen: Interventional Data Generation for Robust and Data-Efficient Robot Imitation Learning
Ryan Hoque, Ajay Mandlekar, Caelan Garrett, Ken Goldberg, Dieter Fox
Continual Imitation Learning for Prosthetic Limbs
Sharmita Dey, Benjamin Paassen, Sarath Ravindran Nair, Sabri Boughorbel, Arndt F. Schilling
Overcoming Knowledge Barriers: Online Imitation Learning from Observation with Pretrained World Models
Xingyuan Zhang, Philip Becker-Ehmck, Patrick van der Smagt, Maximilian Karl
MRIC: Model-Based Reinforcement-Imitation Learning with Mixture-of-Codebooks for Autonomous Driving Simulation
Baotian He, Yibing Li
IDIL: Imitation Learning of Intent-Driven Expert Behavior
Sangwon Seo, Vaibhav Unhelkar
Distilling Privileged Information for Dubins Traveling Salesman Problems with Neighborhoods
Min Kyu Shin, Su-Jeong Park, Seung-Keol Ryu, Heeyeon Kim, Han-Lim Choi
Leveraging Pretrained Latent Representations for Few-Shot Imitation Learning on a Dexterous Robotic Hand
Davide Liconti, Yasunori Toshimitsu, Robert Katzschmann