Visual Reinforcement Learning
Visual reinforcement learning (VRL) aims to train agents to make decisions and control actions directly from visual inputs, a challenging problem due to the high dimensionality and complexity of visual data. Current research heavily focuses on improving sample efficiency and generalization capabilities, employing techniques like contrastive learning, knowledge distillation, and data augmentation within various model architectures including convolutional neural networks and transformers. These advancements are crucial for enabling robots and autonomous systems to learn complex tasks from visual experience, impacting fields like robotics, autonomous driving, and video game AI.
Papers
TACO: Temporal Latent Action-Driven Contrastive Loss for Visual Reinforcement Learning
Ruijie Zheng, Xiyao Wang, Yanchao Sun, Shuang Ma, Jieyu Zhao, Huazhe Xu, Hal Daumé, Furong Huang
Learning from Visual Observation via Offline Pretrained State-to-Go Transformer
Bohan Zhou, Ke Li, Jiechuan Jiang, Zongqing Lu