Absolute Stance Bias
Absolute stance bias refers to systematic errors in machine learning models stemming from skewed training data or algorithmic design, leading to unfair or inaccurate predictions across different groups or contexts. Current research focuses on quantifying and mitigating these biases in various models, including large language models (LLMs), machine translation systems, and image recognition algorithms, often employing techniques like counterfactual fairness, reinforcement learning, and bias-aware evaluation metrics. Understanding and addressing absolute stance bias is crucial for ensuring fairness, reliability, and trustworthiness in AI systems across diverse applications, from healthcare and finance to social media and education.
Papers
Hey GPT, Can You be More Racist? Analysis from Crowdsourced Attempts to Elicit Biased Content from Generative AI
Hangzhi Guo, Pranav Narayanan Venkit, Eunchae Jang, Mukund Srinath, Wenbo Zhang, Bonam Mingole, Vipul Gupta, Kush R. Varshney, S. Shyam Sundar, Amulya Yadav
Exploring Social Desirability Response Bias in Large Language Models: Evidence from GPT-4 Simulations
Sanguk Lee, Kai-Qi Yang, Tai-Quan Peng, Ruth Heo, Hui Liu
CalibraEval: Calibrating Prediction Distribution to Mitigate Selection Bias in LLMs-as-Judges
Haitao Li, Junjie Chen, Qingyao Ai, Zhumin Chu, Yujia Zhou, Qian Dong, Yiqun Liu