Reinforcement Learning
Reinforcement learning (RL) focuses on training agents to make optimal decisions in an environment by learning through trial and error, aiming to maximize cumulative rewards. Current research emphasizes improving RL's efficiency and robustness, particularly in areas like human-in-the-loop training (e.g., using human feedback to refine models), handling uncertainty and sparse rewards, and scaling to complex tasks (e.g., robotics, autonomous driving). Prominent approaches involve various policy gradient methods, Monte Carlo Tree Search, and the integration of large language models for improved decision-making and task decomposition. These advancements are driving progress in diverse fields, including robotics, game playing, and the development of more human-aligned AI systems.
Papers
When to Trust Your Data: Enhancing Dyna-Style Model-Based Reinforcement Learning With Data Filter
Yansong Li, Zeyu Dong, Ertai Luo, Yu Wu, Shuo Wu, Shuo Han
Sample-Efficient Reinforcement Learning with Temporal Logic Objectives: Leveraging the Task Specification to Guide Exploration
Yiannis Kantaros, Jun Wang
Mitigating Suboptimality of Deterministic Policy Gradients in Complex Q-functions
Ayush Jain, Norio Kosaka, Xinhu Li, Kyung-Min Kim, Erdem Bıyık, Joseph J. Lim
Zero-shot Model-based Reinforcement Learning using Large Language Models
Abdelhakim Benechehab, Youssef Attia El Hili, Ambroise Odonnat, Oussama Zekri, Albert Thomas, Giuseppe Paolo, Maurizio Filippone, Ievgen Redko, Balázs Kégl
BlendRL: A Framework for Merging Symbolic and Neural Policy Learning
Hikaru Shindo, Quentin Delfosse, Devendra Singh Dhami, Kristian Kersting
Safety Filtering While Training: Improving the Performance and Sample Efficiency of Reinforcement Learning Agents
Federico Pizarro Bejarano, Lukas Brunke, Angela P. Schoellig
Improve Value Estimation of Q Function and Reshape Reward with Monte Carlo Tree Search
Jiamian Li
SDS -- See it, Do it, Sorted: Quadruped Skill Synthesis from Single Video Demonstration
Jeffrey Li, Maria Stamatopoulou, Dimitrios Kanoulas
Meta-DT: Offline Meta-RL as Conditional Sequence Modeling with World Model Disentanglement
Zhi Wang, Li Zhang, Wenhao Wu, Yuanheng Zhu, Dongbin Zhao, Chunlin Chen
Visual Manipulation with Legs
Xialin He, Chengjing Yuan, Wenxuan Zhou, Ruihan Yang, David Held, Xiaolong Wang
Diffusion-Based Offline RL for Improved Decision-Making in Augmented ARC Task
Yunho Kim, Jaehyun Park, Heejun Kim, Sejin Kim, Byung-Jun Lee, Sundong Kim
ILAEDA: An Imitation Learning Based Approach for Automatic Exploratory Data Analysis
Abhijit Manatkar, Devarsh Patel, Hima Patel, Naresh Manwani
Unveiling Options with Neural Decomposition
Mahdi Alikhasi, Levi H. S. Lelis
Disentangled Unsupervised Skill Discovery for Efficient Hierarchical Reinforcement Learning
Jiaheng Hu, Zizhao Wang, Peter Stone, Roberto Martín-Martín
Bayes Adaptive Monte Carlo Tree Search for Offline Model-based Reinforcement Learning
Jiayu Chen, Wentse Chen, Jeff Schneider
Multi-objective Reinforcement Learning: A Tool for Pluralistic Alignment
Peter Vamplew, Conor F Hayes, Cameron Foale, Richard Dazeley, Hadassah Harland
Reinforcement Learning Based Bidding Framework with High-dimensional Bids in Power Markets
Jinyu Liu, Hongye Guo, Yun Li, Qinghu Tang, Fuquan Huang, Tunan Chen, Haiwang Zhong, Qixin Chen
Action Gaps and Advantages in Continuous-Time Distributional Reinforcement Learning
Harley Wiltzer, Marc G. Bellemare, David Meger, Patrick Shafto, Yash Jhaveri
Improving the Language Understanding Capabilities of Large Language Models Using Reinforcement Learning
Bokai Hu, Sai Ashish Somayajula, Xin Pan, Zihan Huang, Pengtao Xie
Transforming Game Play: A Comparative Study of DCQN and DTQN Architectures in Reinforcement Learning
William A. Stigall