Active Learning
Active learning is a machine learning paradigm focused on optimizing data labeling efficiency by strategically selecting the most informative samples for annotation from a larger unlabeled pool. Current research emphasizes developing novel acquisition functions and data pruning strategies to reduce computational costs associated with large datasets, exploring the integration of active learning with various model architectures (including deep neural networks, Gaussian processes, and language models), and addressing challenges like privacy preservation and handling open-set noise. This approach holds significant promise for reducing the substantial cost and effort of data labeling in diverse fields, ranging from image classification and natural language processing to materials science and healthcare.
Papers
ChemSpaceAL: An Efficient Active Learning Methodology Applied to Protein-Specific Molecular Generation
Gregory W. Kyro, Anton Morgunov, Rafael I. Brent, Victor S. Batista
Stream-based Active Learning by Exploiting Temporal Properties in Perception with Temporal Predicted Loss
Sebastian Schmidt, Stephan Günnemann
Learning Objective-Specific Active Learning Strategies with Attentive Neural Processes
Tim Bakker, Herke van Hoof, Max Welling
Maturity-Aware Active Learning for Semantic Segmentation with Hierarchically-Adaptive Sample Assessment
Amirsaeed Yazdani, Xuelu Li, Vishal Monga
Distributionally Robust Statistical Verification with Imprecise Neural Networks
Souradeep Dutta, Michele Caprio, Vivian Lin, Matthew Cleaveland, Kuk Jin Jang, Ivan Ruchkin, Oleg Sokolsky, Insup Lee