Bias Mitigation
Bias mitigation in machine learning aims to create fairer and more equitable algorithms by addressing biases stemming from training data and model architectures. Current research focuses on developing and evaluating various bias mitigation techniques, including data augmentation strategies (like mixup and proximity sampling), adversarial training methods, and post-processing approaches such as channel pruning and dropout. These efforts span diverse applications, from computer vision and natural language processing to medical image analysis and recommender systems, highlighting the broad significance of this field for ensuring responsible and ethical AI development. The ultimate goal is to improve model fairness without sacrificing accuracy or utility, leading to more equitable outcomes across different demographic groups.
Papers
Data Augmentation via Subgroup Mixup for Improving Fairness
Madeline Navarro, Camille Little, Genevera I. Allen, Santiago Segarra
Mitigating Group Bias in Federated Learning for Heterogeneous Devices
Khotso Selialia, Yasra Chandio, Fatima M. Anwar
Bayesian uncertainty-weighted loss for improved generalisability on polyp segmentation task
Rebecca S. Stone, Pedro E. Chavarrias-Solano, Andrew J. Bulpitt, David C. Hogg, Sharib Ali