Moral Bias
Moral bias in artificial intelligence, particularly large language models (LLMs), is a burgeoning research area focused on identifying and mitigating unfair or discriminatory outputs stemming from biases in training data and model architecture. Current research investigates how LLMs reflect and amplify existing societal biases across different languages and cultures, employing various techniques including analyzing model responses to ethical dilemmas and comparing them to human judgments from diverse groups. Understanding and addressing these biases is crucial for ensuring fairness, transparency, and ethical deployment of AI systems across various applications, ranging from content moderation to decision-support tools.
Papers
November 14, 2024
October 17, 2024
July 21, 2024
April 29, 2024
February 16, 2024
December 8, 2023
November 27, 2023
November 20, 2023
August 1, 2023
July 28, 2023
June 2, 2023
November 14, 2022
September 24, 2022
April 6, 2022