Constrained Reinforcement Learning
Constrained Reinforcement Learning (CRL) addresses the challenge of training agents to maximize rewards while simultaneously satisfying safety or resource constraints, crucial for deploying RL in real-world applications. Current research focuses on developing efficient algorithms, such as primal-dual methods, penalty function methods, and those incorporating techniques like log barrier functions or posterior sampling, often within model-based or model-free frameworks. These advancements improve the safety and reliability of RL agents across diverse domains, including robotics, resource allocation, and safe navigation, by ensuring learned policies adhere to critical operational limitations. The resulting improvements in robustness and safety are significant for transitioning RL from simulated to real-world deployments.
Papers
Provably Efficient Exploration in Constrained Reinforcement Learning:Posterior Sampling Is All You Need
Danil Provodin, Pratik Gajane, Mykola Pechenizkiy, Maurits Kaptein
Evaluation of Constrained Reinforcement Learning Algorithms for Legged Locomotion
Joonho Lee, Lukas Schroth, Victor Klemm, Marko Bjelonic, Alexander Reske, Marco Hutter