Positive Unlabeled Data

Positive-unlabeled (PU) learning tackles the challenge of training classifiers using datasets containing only positively labeled examples and unlabeled data, a common scenario in many applications. Current research focuses on improving classifier performance under various sampling schemes (e.g., case-control vs. single-sample) and addressing biases in the data, employing techniques like asymmetric loss functions, variational autoencoders, and logistic regression approaches. These advancements aim to enhance the accuracy and robustness of PU learning methods, particularly for complex scenarios such as detecting AI-generated text or improving image quality assessment, where obtaining fully labeled datasets is impractical or costly.

Papers