Stereotypical Bias

Stereotypical bias in artificial intelligence, particularly in large language models (LLMs) and text-to-image generators, focuses on identifying and mitigating the perpetuation of harmful societal biases encoded within these systems. Current research investigates bias across various demographic categories (gender, race, ethnicity, etc.) using diverse datasets and evaluation metrics, often employing techniques like reinforcement learning and counterfactual data generation to debias models. This work is crucial for ensuring fairness and equity in AI applications, impacting not only the development of responsible AI but also our understanding of how biases are learned and propagated through machine learning systems.

Papers