Diffusion Policy
Diffusion policies leverage the power of diffusion models to generate actions for complex tasks, primarily aiming to improve the robustness, efficiency, and generalization capabilities of reinforcement learning agents and robotic controllers. Current research focuses on refining algorithms like Diffusion Policy Policy Optimization (DPPO) and exploring architectures such as Mixture of Experts (MoE) to enhance multi-task learning and reduce computational costs, often incorporating techniques from reinforcement learning and imitation learning. This approach holds significant promise for advancing robotics, particularly in areas like manipulation, locomotion, and navigation, by enabling more adaptable and data-efficient learning of complex behaviors.
Papers
Planning-Guided Diffusion Policy Learning for Generalizable Contact-Rich Bimanual Manipulation
Xuanlin Li, Tong Zhao, Xinghao Zhu, Jiuguang Wang, Tao Pang, Kuan Fang
Diffusion Implicit Policy for Unpaired Scene-aware Motion Synthesis
Jingyu Gong, Chong Zhang, Fengqi Liu, Ke Fan, Qianyu Zhou, Xin Tan, Zhizhong Zhang, Yuan Xie, Lizhuang Ma