Debiasing Method
Debiasing methods aim to mitigate biases learned by machine learning models, particularly large language models (LLMs), from skewed training data, improving fairness and generalizability. Current research focuses on techniques like data augmentation, weight masking, and prompt engineering, often applied to various architectures including BERT and other transformer-based models, as well as employing causal inference and unlearning approaches. Effective debiasing is crucial for ensuring fairness in AI applications across diverse domains, ranging from healthcare and criminal justice to natural language processing tasks, and is a significant area of ongoing investigation within the broader field of AI ethics.
Papers
Debiasing Methods for Fairer Neural Models in Vision and Language Research: A Survey
Otávio Parraga, Martin D. More, Christian M. Oliveira, Nathan S. Gavenski, Lucas S. Kupssinskü, Adilson Medronha, Luis V. Moura, Gabriel S. Simões, Rodrigo C. Barros
ADEPT: A DEbiasing PrompT Framework
Ke Yang, Charles Yu, Yi Fung, Manling Li, Heng Ji