Biased Behavior
Biased behavior in artificial intelligence, particularly large language models (LLMs) and other machine learning systems, is a significant area of research focusing on identifying, mitigating, and understanding the sources of such biases. Current efforts utilize various techniques, including Bayesian methods for bias removal, multitask learning to disentangle dialect from bias, and the development of detectors (guardrails) trained on synthetic data to flag problematic outputs. This research is crucial for ensuring fairness and equity in AI applications, impacting fields ranging from news consumption and social media to healthcare and loan applications, and promoting the development of more trustworthy and responsible AI systems.
Papers
Towards A Reliable Ground-Truth For Biased Language Detection
Timo Spinde, David Krieger, Manuel Plank, Bela Gipp
Do You Think It's Biased? How To Ask For The Perception Of Media Bias
Timo Spinde, Christina Kreuter, Wolfgang Gaissmaier, Felix Hamborg, Bela Gipp, Helge Giese
Identification of Biased Terms in News Articles by Comparison of Outlet-specific Word Embeddings
Timo Spinde, Lada Rudnitckaia, Felix Hamborg, Bela Gipp