Active Learning
Active learning is a machine learning paradigm focused on optimizing data labeling efficiency by strategically selecting the most informative samples for annotation from a larger unlabeled pool. Current research emphasizes developing novel acquisition functions and data pruning strategies to reduce computational costs associated with large datasets, exploring the integration of active learning with various model architectures (including deep neural networks, Gaussian processes, and language models), and addressing challenges like privacy preservation and handling open-set noise. This approach holds significant promise for reducing the substantial cost and effort of data labeling in diverse fields, ranging from image classification and natural language processing to materials science and healthcare.
Papers
Minority Class Oriented Active Learning for Imbalanced Datasets
Umang Aggarwal, Adrian Popescu, Céline Hudelot
Active Learning Over Multiple Domains in Natural Language Tasks
Shayne Longpre, Julia Reisler, Edward Greg Huang, Yi Lu, Andrew Frank, Nikhil Ramesh, Chris DuBois
Federated Active Learning (F-AL): an Efficient Annotation Strategy for Federated Learning
Jin-Hyun Ahn, Kyungsang Kim, Jeongwan Koh, Quanzheng Li
How Low Can We Go? Pixel Annotation for Semantic Segmentation
Daniel Kigli, Ariel Shamir, Shai Avidan
Little Help Makes a Big Difference: Leveraging Active Learning to Improve Unsupervised Time Series Anomaly Detection
Hamza Bodor, Thai V. Hoang, Zonghua Zhang
Cold Start Active Learning Strategies in the Context of Imbalanced Classification
Etienne Brangbour, Pierrick Bruneau, Thomas Tamisier, Stéphane Marchand-Maillet
Keeping Deep Lithography Simulators Updated: Global-Local Shape-Based Novelty Detection and Active Learning
Hao-Chiang Shao, Hsing-Lei Ping, Kuo-shiuan Chen, Weng-Tai Su, Chia-Wen Lin, Shao-Yun Fang, Pin-Yian Tsai, Yan-Hsiu Liu
Active Learning Polynomial Threshold Functions
Omri Ben-Eliezer, Max Hopkins, Chutong Yang, Hantao Yu