Learning Reward
Learning reward functions in reinforcement learning aims to automatically define what constitutes desirable behavior for an AI agent, eliminating the need for manual reward engineering. Current research focuses on methods that learn rewards from various sources, including expert demonstrations (e.g., using diffusion models or ELO ratings), human feedback (incorporating models of human rationality), and even implicitly through guiding agents towards global performance metrics. These advancements are crucial for building more robust and reliable AI systems, particularly in complex, real-world scenarios where manually designing reward functions is impractical or impossible.
Papers
September 5, 2024
August 16, 2024
February 7, 2024
December 21, 2023
May 3, 2023
April 13, 2023
March 16, 2023