Preference Learning
Preference learning aims to align artificial intelligence models, particularly large language models, with human preferences by learning from human feedback on model outputs. Current research focuses on developing efficient algorithms, such as direct preference optimization and reinforcement learning from human feedback, often incorporating advanced model architectures like diffusion models and variational autoencoders to handle complex preference structures, including intransitivity. This field is crucial for building trustworthy and beneficial AI systems, improving their performance on various tasks and ensuring alignment with human values in diverse applications ranging from robotics to natural language processing.
Papers
March 13, 2024
March 8, 2024
March 5, 2024
February 22, 2024
February 18, 2024
February 15, 2024
February 2, 2024
January 26, 2024
January 11, 2024
December 28, 2023
December 27, 2023
December 14, 2023
December 13, 2023
December 5, 2023
December 1, 2023
November 23, 2023
November 21, 2023
October 24, 2023