Deterministic Policy Gradient
Deterministic Policy Gradient (DPG) methods aim to find optimal control policies for continuous systems by directly optimizing a deterministic policy, rather than learning a probability distribution over actions. Current research focuses on improving DPG's accuracy and efficiency, addressing challenges like imprecise gradient estimations through techniques such as zeroth-order approximations and primal-dual methods, and enhancing sample efficiency via data augmentation and adaptive regularization. These advancements are significant for applications in robotics, control systems, and other domains requiring efficient and robust control of continuous-state systems, particularly in constrained or non-stationary environments.
Papers
October 15, 2024
September 2, 2024
August 19, 2024
July 13, 2024
July 5, 2024
May 22, 2024
April 19, 2024
May 31, 2023
May 24, 2023
February 17, 2023
February 5, 2023
October 15, 2022
July 27, 2022
March 25, 2022