Reinforcement Learning
Reinforcement learning (RL) focuses on training agents to make optimal decisions in an environment by learning through trial and error, aiming to maximize cumulative rewards. Current research emphasizes improving RL's efficiency and robustness, particularly in areas like human-in-the-loop training (e.g., using human feedback to refine models), handling uncertainty and sparse rewards, and scaling to complex tasks (e.g., robotics, autonomous driving). Prominent approaches involve various policy gradient methods, Monte Carlo Tree Search, and the integration of large language models for improved decision-making and task decomposition. These advancements are driving progress in diverse fields, including robotics, game playing, and the development of more human-aligned AI systems.
Papers
On-orbit Servicing for Spacecraft Collision Avoidance With Autonomous Decision Making
Susmitha Patnala, Adam Abdin
Reinforcement Learning for Finite Space Mean-Field Type Games
Kai Shao, Jiacheng Shen, Chijie An, Mathieu Laurière
A random measure approach to reinforcement learning in continuous time
Christian Bender, Nguyen Tran Thuan
Dynamic Obstacle Avoidance through Uncertainty-Based Adaptive Planning with Diffusion
Vineet Punyamoorty, Pascal Jutras-Dubé, Ruqi Zhang, Vaneet Aggarwal, Damon Conover, Aniket Bera
Behavior evolution-inspired approach to walking gait reinforcement training for quadruped robots
Yu Wang, Wenchuan Jia, Yi Sun, Dong He
OffRIPP: Offline RL-based Informative Path Planning
Srikar Babu Gadipudi, Srujan Deolasee, Siva Kailas, Wenhao Luo, Katia Sycara, Woojun Kim
Symbolic State Partition for Reinforcement Learning
Mohsen Ghaffari, Mahsa Varshosaz, Einar Broch Johnsen, Andrzej Wąsowski
Offline and Distributional Reinforcement Learning for Radio Resource Management
Eslam Eldeeb, Hirley Alves
Optimized Monte Carlo Tree Search for Enhanced Decision Making in the FrozenLake Environment
Esteban Aldana Guerra
From Goal-Conditioned to Language-Conditioned Agents via Vision-Language Models
Theo Cachet, Christopher R. Dance, Olivier Sigaud
Stage-Wise Reward Shaping for Acrobatic Robots: A Constrained Multi-Objective Reinforcement Learning Approach
Dohyeong Kim, Hyeokjin Kwon, Junseok Kim, Gunmin Lee, Songhwai Oh
Development and Validation of Heparin Dosing Policies Using an Offline Reinforcement Learning Algorithm
Yooseok Lim, Inbeom Park, Sujee Lee
Walking with Terrain Reconstruction: Learning to Traverse Risky Sparse Footholds
Ruiqi Yu, Qianshi Wang, Yizhen Wang, Zhicheng Wang, Jun Wu, Qiuguo Zhu
SurgIRL: Towards Life-Long Learning for Surgical Automation by Incremental Reinforcement Learning
Yun-Jie Ho, Zih-Yun Chiu, Yuheng Zhi, Michael C. Yip
Physics Enhanced Residual Policy Learning (PERPL) for safety cruising in mixed traffic platooning under actuator and communication delay
Keke Long, Haotian Shi, Yang Zhou, Xiaopeng Li
Learning Diverse Robot Striking Motions with Diffusion Models and Kinematically Constrained Gradient Guidance
Kin Man Lee, Sean Ye, Qingyu Xiao, Zixuan Wu, Zulfiqar Zaidi, David B. D'Ambrosio, Pannag R. Sanketi, Matthew Gombolay
CANDERE-COACH: Reinforcement Learning from Noisy Feedback
Yuxuan Li, Srijita Das, Matthew E. Taylor
Intelligent Routing Algorithm over SDN: Reusable Reinforcement Learning Approach
Wang Wumian, Sajal Saha, Anwar Haque, Greg Sidebottom
SPformer: A Transformer Based DRL Decision Making Method for Connected Automated Vehicles
Ye Han, Lijun Zhang, Dejian Meng, Xingyu Hu, Yixia Lu
Deep Reinforcement Learning-based Obstacle Avoidance for Robot Movement in Warehouse Environments
Keqin Li, Jiajing Chen, Denzhi Yu, Tao Dajun, Xinyu Qiu, Lian Jieting, Sun Baiwei, Zhang Shengyuan, Zhenyu Wan, Ran Ji, Bo Hong, Fanghao Ni