Annotation Bias

Annotation bias, the systematic error introduced by human labelers in machine learning datasets, significantly impacts the fairness and accuracy of trained models. Current research focuses on identifying and mitigating these biases, particularly in natural language processing and medical image segmentation, using techniques like multi-variable causal inference and transformer-based models to disentangle annotator preferences from stochastic errors. Understanding and addressing annotation bias is crucial for developing robust and reliable AI systems across various domains, improving the trustworthiness and generalizability of machine learning models.

Papers