Label Noise
Label noise, the presence of incorrect labels in training datasets, significantly hinders the performance and robustness of machine learning models. Current research focuses on developing methods to mitigate this issue, exploring techniques like loss function modifications, sample selection strategies (e.g., identifying and removing or down-weighting noisy samples), and the use of robust algorithms such as those based on nearest neighbors or contrastive learning, often applied within deep neural networks or gradient boosted decision trees. Addressing label noise is crucial for improving the reliability and generalizability of machine learning models across various applications, from medical image analysis to natural language processing, and is driving the development of new benchmark datasets and evaluation metrics.
Papers
Best Transition Matrix Esitimation or Best Label Noise Robustness Classifier? Two Possible Methods to Enhance the Performance of T-revision
Haixu Liu, Zerui Tao, Naihui Zhang, Sixing Liu
An Inclusive Theoretical Framework of Robust Supervised Contrastive Loss against Label Noise
Jingyi Cui, Yi-Ge Zhang, Hengyu Liu, Yisen Wang
Selfish Evolution: Making Discoveries in Extreme Label Noise with the Help of Overfitting Dynamics
Nima Sedaghat, Tanawan Chatchadanoraset, Colin Orion Chandler, Ashish Mahabal, Maryam Eslami
Improving Resistance to Noisy Label Fitting by Reweighting Gradient in SAM
Hoang-Chau Luong, Thuc Nguyen-Quang, Minh-Triet Tran