Stereotype Content
Stereotype content research investigates how biases and stereotypes are represented and perpetuated within large language models (LLMs) and other AI systems, aiming to understand and mitigate their harmful societal impact. Current research focuses on identifying and quantifying these biases across various modalities (text, images), languages, and demographic groups, often employing techniques like adversarial attacks and explainable AI methods to analyze model behavior and develop mitigation strategies. This work is crucial for ensuring fairness and equity in AI applications, impacting fields ranging from education and healthcare to hiring and criminal justice, by promoting the development of less biased and more responsible AI systems.
Papers
May 29, 2023
May 26, 2023
May 23, 2023
May 19, 2023
March 28, 2023
February 27, 2023
February 14, 2023
February 3, 2023
January 11, 2023
December 20, 2022
November 21, 2022
November 7, 2022
October 26, 2022
October 11, 2022
August 2, 2022
July 23, 2022
June 23, 2022
May 27, 2022
May 22, 2022
April 4, 2022