Trajectory Preference

Trajectory preference research focuses on aligning reinforcement learning (RL) agents' behavior with human preferences, expressed through comparisons of entire agent trajectories rather than instantaneous rewards. Current research emphasizes efficient exploration and learning from limited human feedback, employing methods like dynamic policy fusion, preference-guided policy optimization, and inverse reinforcement learning to infer underlying reward functions from trajectory preferences. This field is crucial for developing safe and user-friendly RL agents in complex applications, such as autonomous driving and robotics, where designing explicit reward functions is difficult or impossible.

Papers