Human Label Variation
Human label variation (HLV), the inherent disagreement among human annotators labeling the same data, is a significant challenge in natural language processing (NLP). Current research focuses on understanding the sources of HLV (e.g., ambiguity, differing interpretations), developing methods to model and predict this variation (often employing multi-head models or techniques inspired by recommender systems), and improving model calibration by accounting for HLV rather than treating it as noise. Addressing HLV is crucial for building more robust, trustworthy, and fair NLP systems, impacting the reliability of benchmarks and the development of more accurate and interpretable models.
Papers
Variationist: Exploring Multifaceted Variation and Bias in Written Language Data
Alan Ramponi, Camilla Casula, Stefano Menini
"Seeing the Big through the Small": Can LLMs Approximate Human Judgment Distributions on NLI from a Few Explanations?
Beiduo Chen, Xinpeng Wang, Siyao Peng, Robert Litschko, Anna Korhonen, Barbara Plank