Left Leaning

"Left leaning," in the context of recent research, refers broadly to the observed biases in various machine learning models, particularly large language models (LLMs), towards left-leaning political viewpoints. Current research focuses on quantifying and understanding these biases, investigating their origins (e.g., in training data or model architecture), and developing methods to mitigate them. This is significant because the increasing use of LLMs in information retrieval and decision-making necessitates unbiased outputs, impacting fairness, accuracy, and the prevention of filter bubbles.

Papers