Preference Rating
Preference rating research focuses on efficiently and accurately capturing and utilizing human preferences, particularly in the context of ranking and comparing outputs from complex systems like large language models (LLMs) and reinforcement learning agents. Current research emphasizes developing robust methods for preference elicitation, including incorporating response times and handling incomplete or noisy data, often employing techniques like linear bandits, contrastive learning, and various ranking algorithms (e.g., DPO, spectral methods). These advancements are crucial for improving the alignment of AI systems with human values and preferences, leading to more reliable and trustworthy AI applications across diverse fields.
Papers
November 15, 2024
October 11, 2024
October 7, 2024
October 5, 2024
September 9, 2024
July 30, 2024
July 2, 2024
June 22, 2024
June 3, 2024
May 29, 2024
March 28, 2024
March 12, 2024
February 2, 2024
September 28, 2023
August 5, 2023
June 30, 2023
June 6, 2023
May 16, 2023
May 15, 2023
October 11, 2022