Token Bias
Token bias in large language models (LLMs) refers to the disproportionate influence of specific tokens (sub-word units) on model outputs, leading to inaccurate or unfair predictions. Current research focuses on identifying and mitigating this bias through various techniques, including recalibrating automated evaluators, developing unbiased tokenization algorithms, and employing methods like contrastive clustering to disentangle causal and correlational relationships between tokens and model predictions. Addressing token bias is crucial for improving the reliability, fairness, and generalizability of LLMs across diverse applications, ranging from sentiment analysis and recommendation systems to fact-checking and toxic language detection.
Papers
November 17, 2024
November 1, 2024
October 15, 2024
October 11, 2024
July 5, 2024
June 24, 2024
June 18, 2024
June 16, 2024
June 3, 2024
May 27, 2024
May 22, 2024
April 27, 2024
February 22, 2024
January 27, 2024
September 7, 2023
July 19, 2022
April 15, 2022
December 23, 2021