State Action Pair
State-action pairs represent the fundamental units of interaction in reinforcement learning (RL), encompassing the agent's state and chosen action at a given time step. Current research focuses on improving the representation and utilization of state-action pairs, particularly through advanced model architectures like diffusion models and ensemble methods, and algorithms such as those employing bisimulation or contrastive learning to enhance sample efficiency and robustness. This work is crucial for advancing RL's capabilities in complex environments, particularly in offline settings and safety-critical applications, by enabling more accurate value function estimation, improved policy learning, and better generalization.
Papers
Diverse Policies Recovering via Pointwise Mutual Information Weighted Imitation Learning
Hanlin Yang, Jian Yao, Weiming Liu, Qing Wang, Hanmin Qin, Hansheng Kong, Kirk Tang, Jiechao Xiong, Chao Yu, Kai Li, Junliang Xing, Hongwu Chen, Juchao Zhuo, Qiang Fu, Yang Wei, Haobo Fu
In-Trajectory Inverse Reinforcement Learning: Learn Incrementally Before An Ongoing Trajectory Terminates
Shicheng Liu, Minghui Zhu
The Number of Trials Matters in Infinite-Horizon General-Utility Markov Decision Processes
Pedro P. Santos, Alberto Sardinha, Francisco S. Melo
A novel agent with formal goal-reaching guarantees: an experimental study with a mobile robot
Grigory Yaremenko, Dmitrii Dobriborsci, Roman Zashchitin, Ruben Contreras Maestre, Ngoc Quoc Huy Hoang, Pavel Osinenko