Preference Learning
Preference learning aims to align artificial intelligence models, particularly large language models, with human preferences by learning from human feedback on model outputs. Current research focuses on developing efficient algorithms, such as direct preference optimization and reinforcement learning from human feedback, often incorporating advanced model architectures like diffusion models and variational autoencoders to handle complex preference structures, including intransitivity. This field is crucial for building trustworthy and beneficial AI systems, improving their performance on various tasks and ensuring alignment with human values in diverse applications ranging from robotics to natural language processing.
Papers
September 22, 2023
September 7, 2023
September 6, 2023
August 18, 2023
June 9, 2023
June 8, 2023
May 30, 2023
May 24, 2023
May 18, 2023
April 12, 2023
April 7, 2023
March 1, 2023
December 6, 2022
September 11, 2022
July 25, 2022
July 7, 2022
June 10, 2022
June 9, 2022
May 16, 2022