Robust Reward
Robust reward design in reinforcement learning aims to create reward functions that reliably guide agents towards desired behaviors, even in the presence of noise, ambiguity, or unexpected situations. Current research focuses on improving reward model robustness through techniques like noise-resilient reward functions (e.g., using binary mutual information), causal frameworks to filter out irrelevant artifacts, and methods that learn from noisy or suboptimal demonstrations. These advancements are crucial for deploying reinforcement learning agents in real-world settings where perfect reward signals are unavailable, impacting fields like robotics, natural language processing, and autonomous systems.
Papers
November 13, 2024
September 24, 2024
September 20, 2024
August 27, 2024
August 22, 2024
July 24, 2024
June 21, 2024
June 10, 2024
June 7, 2024
May 8, 2024
April 10, 2024
November 22, 2023
October 26, 2023
October 23, 2023
October 11, 2023
June 6, 2023
May 17, 2023
January 2, 2023
July 28, 2022