Implicit Bias
Implicit bias refers to unintended, often subtle, biases embedded within machine learning models, stemming from biases present in their training data. Current research focuses on detecting and mitigating these biases in various model architectures, particularly large language models (LLMs) and deep neural networks, using techniques like prompt engineering, fine-tuning, and Bayesian methods. Understanding and addressing implicit bias is crucial for ensuring fairness and equity in AI applications, impacting fields ranging from healthcare and criminal justice to education and hiring. The development of robust bias detection and mitigation strategies is a central goal of ongoing research.
Papers
February 9, 2022
February 5, 2022
February 3, 2022
February 2, 2022
January 27, 2022
January 22, 2022
January 12, 2022
January 3, 2022
December 23, 2021
December 13, 2021
December 10, 2021
November 16, 2021