Constrained Reinforcement Learning
Constrained Reinforcement Learning (CRL) addresses the challenge of training agents to maximize rewards while simultaneously satisfying safety or resource constraints, crucial for deploying RL in real-world applications. Current research focuses on developing efficient algorithms, such as primal-dual methods, penalty function methods, and those incorporating techniques like log barrier functions or posterior sampling, often within model-based or model-free frameworks. These advancements improve the safety and reliability of RL agents across diverse domains, including robotics, resource allocation, and safe navigation, by ensuring learned policies adhere to critical operational limitations. The resulting improvements in robustness and safety are significant for transitioning RL from simulated to real-world deployments.
Papers
Probabilistic Satisfaction of Temporal Logic Constraints in Reinforcement Learning via Adaptive Policy-Switching
Xiaoshan Lin, Sadık Bera Yüksel, Yasin Yazıcıoğlu, Derya Aksaray
Offline Inverse Constrained Reinforcement Learning for Safe-Critical Decision Making in Healthcare
Nan Fang, Guiliang Liu, Wei Gong