Bias Evaluation
Bias evaluation in machine learning focuses on identifying and quantifying unfair biases in models' outputs, aiming to promote fairness and mitigate discriminatory outcomes. Current research emphasizes developing new metrics and benchmarks to assess bias across diverse model architectures, including large language models and computer vision systems, often employing techniques like counterfactual analysis and probing methods to detect subtle biases. This work is crucial for ensuring the responsible development and deployment of AI systems, impacting fields ranging from healthcare and criminal justice to social media and autonomous driving, where biased algorithms can have significant societal consequences.
Papers
CHBias: Bias Evaluation and Mitigation of Chinese Conversational Language Models
Jiaxu Zhao, Meng Fang, Zijing Shi, Yitong Li, Ling Chen, Mykola Pechenizkiy
Comparing Biases and the Impact of Multilingual Training across Multiple Languages
Sharon Levy, Neha Anna John, Ling Liu, Yogarshi Vyas, Jie Ma, Yoshinari Fujinuma, Miguel Ballesteros, Vittorio Castelli, Dan Roth