Robot Policy
Robot policy research focuses on developing robust and adaptable control algorithms enabling robots to perform complex tasks in diverse and unpredictable environments. Current efforts concentrate on improving policy generalization through techniques like large-scale reinforcement learning fine-tuning, leveraging pre-trained vision-language models for improved reasoning and failure detection, and incorporating human preferences and feedback for personalization. These advancements are crucial for deploying reliable and efficient robots in real-world applications, ranging from assistive robotics to industrial automation.
Papers
ACDC: Automated Creation of Digital Cousins for Robust Policy Learning
Tianyuan Dai, Josiah Wong, Yunfan Jiang, Chen Wang, Cem Gokmen, Ruohan Zhang, Jiajun Wu, Li Fei-Fei
VIRT: Vision Instructed Transformer for Robotic Manipulation
Zhuoling Li, Liangliang Ren, Jinrong Yang, Yong Zhao, Xiaoyang Wu, Zhenhua Xu, Xiang Bai, Hengshuang Zhao
Grounding Robot Policies with Visuomotor Language Guidance
Arthur Bucker, Pablo Ortega, Jonathan Francis, Jean Oh