Temporal Difference
Temporal difference (TD) learning is a core reinforcement learning method aiming to estimate the value of states or state-action pairs by bootstrapping from predictions of future values. Current research focuses on improving TD's efficiency and stability, particularly when combined with deep neural networks, by addressing issues like variance reduction, handling uncertainty, and optimizing algorithm parameters (e.g., step size, target network updates). These advancements are significant for enhancing the performance and robustness of reinforcement learning agents across various applications, from robotics and game playing to more complex control problems and even supervised learning tasks.
Papers
November 6, 2024
August 5, 2024
June 22, 2024
May 27, 2024
May 26, 2024
May 22, 2024
May 7, 2024
April 23, 2024
March 15, 2024
December 8, 2023
September 29, 2023
August 2, 2023
July 10, 2023
June 29, 2023
May 25, 2023
March 13, 2023
February 25, 2023
February 24, 2023
February 20, 2023