Learning Reward

Learning reward functions in reinforcement learning aims to automatically define what constitutes desirable behavior for an AI agent, eliminating the need for manual reward engineering. Current research focuses on methods that learn rewards from various sources, including expert demonstrations (e.g., using diffusion models or ELO ratings), human feedback (incorporating models of human rationality), and even implicitly through guiding agents towards global performance metrics. These advancements are crucial for building more robust and reliable AI systems, particularly in complex, real-world scenarios where manually designing reward functions is impractical or impossible.

Papers