Length Bias

Length bias, a pervasive issue in machine learning, particularly within large language models (LLMs), refers to the tendency of models and evaluation metrics to unfairly favor longer outputs regardless of their actual quality. Current research focuses on identifying and mitigating this bias through various techniques, including reward model calibration, algorithmic modifications like downsampled KL divergence in Direct Preference Optimization (DPO), and data-centric approaches such as length-controlled evaluation metrics. Addressing length bias is crucial for ensuring the reliable evaluation and fair development of LLMs, ultimately leading to more accurate and trustworthy AI systems.

Papers