Return Conditioned Supervised Learning
Return-conditioned supervised learning (RCSL) focuses on training models that predict actions or outputs conditioned on desired future outcomes, such as high rewards or specific target values. Current research explores various model architectures, including diffusion models and decision transformers, and investigates techniques to improve robustness and efficiency, addressing challenges like hyperparameter sensitivity and the impact of environmental stochasticity. This approach holds significant promise for advancing offline reinforcement learning, improving the efficiency of conditional generation tasks in diverse fields, and enabling more controllable and aligned AI systems.
Papers
October 4, 2024
July 2, 2024
June 18, 2024
June 15, 2024
June 3, 2024
April 22, 2024
April 7, 2024
February 8, 2024
February 3, 2024
December 21, 2023
May 19, 2023
April 17, 2023
March 11, 2023
February 20, 2023
February 16, 2023
December 1, 2022
November 29, 2022
November 26, 2022
October 24, 2022
June 2, 2022