Model Bias
Model bias, the tendency of machine learning models to produce unfair or inaccurate predictions for certain subgroups, is a critical area of research aiming to improve model fairness and reliability. Current efforts focus on identifying and mitigating bias through techniques like data augmentation, causal inference methods, and algorithmic adjustments to model architectures such as transformers and neural networks, often employing post-processing or in-processing approaches. Understanding and addressing model bias is crucial for ensuring the responsible development and deployment of AI systems across various applications, impacting fields from healthcare and finance to climate modeling and natural language processing.
Papers
Soft-prompt Tuning for Large Language Models to Evaluate Bias
Jacob-Junqi Tian, David Emerson, Sevil Zanjani Miyandoab, Deval Pandya, Laleh Seyyed-Kalantari, Faiza Khan Khattak
ICON$^2$: Reliably Benchmarking Predictive Inequity in Object Detection
Sruthi Sudhakar, Viraj Prabhu, Olga Russakovsky, Judy Hoffman
How to Construct Perfect and Worse-than-Coin-Flip Spoofing Countermeasures: A Word of Warning on Shortcut Learning
Hye-jin Shim, Rosa González Hautamäki, Md Sahidullah, Tomi Kinnunen
Signal Is Harder To Learn Than Bias: Debiasing with Focal Loss
Moritz Vandenhirtz, Laura Manduchi, Ričards Marcinkevičs, Julia E. Vogt
Cognitive network science reveals bias in GPT-3, ChatGPT, and GPT-4 mirroring math anxiety in high-school students
Katherine Abramski, Salvatore Citraro, Luigi Lombardi, Giulio Rossetti, Massimo Stella
Keeping Up with the Language Models: Systematic Benchmark Extension for Bias Auditing
Ioana Baldini, Chhavi Yadav, Manish Nagireddy, Payel Das, Kush R. Varshney