Offline Policy Learning
Offline policy learning aims to train reinforcement learning agents using pre-collected datasets, avoiding the risks and costs of online interaction. Current research focuses on addressing challenges like data scarcity, distributional shifts, and the presence of suboptimal behaviors in the training data, employing techniques such as importance weighting regularization, skill-based abstractions, deep generative models, and dataset clustering to improve policy performance. These advancements are significant for applications where online learning is impractical or unsafe, such as healthcare, robotics, and safety-critical systems, enabling the development of robust and effective policies from existing data.
Papers
September 22, 2024
August 27, 2024
August 21, 2024
June 5, 2024
February 21, 2024
February 16, 2024
February 14, 2024
January 19, 2024
October 6, 2023
September 18, 2023
March 30, 2023
February 1, 2023
January 30, 2023
January 27, 2023
October 12, 2022
July 26, 2022
November 27, 2021