Preference Rating
Preference rating research focuses on efficiently and accurately capturing and utilizing human preferences, particularly in the context of ranking and comparing outputs from complex systems like large language models (LLMs) and reinforcement learning agents. Current research emphasizes developing robust methods for preference elicitation, including incorporating response times and handling incomplete or noisy data, often employing techniques like linear bandits, contrastive learning, and various ranking algorithms (e.g., DPO, spectral methods). These advancements are crucial for improving the alignment of AI systems with human values and preferences, leading to more reliable and trustworthy AI applications across diverse fields.
Papers
September 22, 2022
February 7, 2022