Partial Label Learning
Partial label learning (PLL) addresses the challenge of training machine learning models using data where each example is annotated with a set of possible labels, only one of which is correct. Current research focuses on improving label disambiguation techniques, often employing deep learning architectures like neural networks and Gaussian processes, and exploring variations such as multi-instance PLL and handling noisy or unreliable labels. These advancements are significant because they enable more efficient and robust model training with less precise annotations, impacting various fields including image classification, natural language processing, and biomedical applications where obtaining perfectly labeled data is difficult or expensive. Furthermore, research is actively addressing issues like class imbalance and out-of-distribution data within the PLL framework.
Papers
G2NetPL: Generic Game-Theoretic Network for Partial-Label Image Classification
Rabab Abdelfattah, Xin Zhang, Mostafa M. Fouda, Xiaofeng Wang, Song Wang
Controller-Guided Partial Label Consistency Regularization with Unlabeled Data
Qian-Wei Wang, Bowen Zhao, Mingyan Zhu, Tianxiang Li, Zimo Liu, Shu-Tao Xia