Offline RL
Offline reinforcement learning (RL) aims to train agents using pre-collected data, avoiding the need for costly and potentially unsafe online interactions. Current research focuses on addressing the challenges of distribution shift (avoiding overestimation of unseen actions) and improving the efficiency and robustness of algorithms, including those leveraging techniques like denoising score matching, implicit Q-learning, and diffusion models. These advancements are significant because they enable the application of RL to real-world scenarios where online data collection is impractical or impossible, impacting fields such as robotics and personalized medicine.
Papers
Contrastive Example-Based Control
Kyle Hatch, Benjamin Eysenbach, Rafael Rafailov, Tianhe Yu, Ruslan Salakhutdinov, Sergey Levine, Chelsea Finn
A Connection between One-Step Regularization and Critic Regularization in Reinforcement Learning
Benjamin Eysenbach, Matthieu Geist, Sergey Levine, Ruslan Salakhutdinov