Conservative Q Learning
Conservative Q-learning (CQL) is an offline reinforcement learning algorithm designed to mitigate the risk of overestimating value functions, a common problem when learning from static datasets. Current research focuses on improving CQL's performance and robustness through techniques like incorporating novel neural network architectures (e.g., Kolmogorov-Arnold Networks), addressing data imbalances, and developing more nuanced approaches to pessimism in value estimation. These advancements are significant because they enhance the reliability and applicability of offline RL in various domains, including robotics, healthcare, and resource management, where online learning is impractical or unsafe.
Papers
October 30, 2024
September 15, 2024
June 20, 2024
June 6, 2024
February 13, 2024
February 3, 2024
December 24, 2023
September 22, 2023
July 6, 2023
June 13, 2023
May 23, 2023
March 9, 2023
March 7, 2023
February 15, 2023
February 14, 2023
January 3, 2023
November 28, 2022
November 2, 2022
October 14, 2022