Biased Behavior
Biased behavior in artificial intelligence, particularly large language models (LLMs) and other machine learning systems, is a significant area of research focusing on identifying, mitigating, and understanding the sources of such biases. Current efforts utilize various techniques, including Bayesian methods for bias removal, multitask learning to disentangle dialect from bias, and the development of detectors (guardrails) trained on synthetic data to flag problematic outputs. This research is crucial for ensuring fairness and equity in AI applications, impacting fields ranging from news consumption and social media to healthcare and loan applications, and promoting the development of more trustworthy and responsible AI systems.
Papers
August 20, 2024
July 15, 2024
July 8, 2024
June 14, 2024
June 13, 2024
May 4, 2024
April 3, 2024
March 8, 2024
March 6, 2024
February 7, 2024
January 11, 2024
December 22, 2023
November 11, 2023
October 12, 2023
October 6, 2023
September 15, 2023
August 28, 2023
May 21, 2023
November 24, 2022