Preference Distribution

Preference distribution research focuses on understanding and modeling how preferences are distributed within a population, aiming to accurately represent diverse viewpoints and avoid biases in decision-making systems. Current research emphasizes robust methods for learning preference distributions from often noisy and incomplete data, employing techniques like reward model distillation, distributional preference learning, and optimal transport-based loss functions within various model architectures, including those based on reinforcement learning and Bayesian updating. This work is crucial for improving the fairness and accuracy of AI systems, particularly in applications like large language model alignment and personalized recommendations, where accurately reflecting diverse user preferences is paramount. The ultimate goal is to develop algorithms that are robust to preference shifts and capable of making equitable decisions even with limited or heterogeneous data.

Papers