Active Learning
Active learning is a machine learning paradigm focused on optimizing data labeling efficiency by strategically selecting the most informative samples for annotation from a larger unlabeled pool. Current research emphasizes developing novel acquisition functions and data pruning strategies to reduce computational costs associated with large datasets, exploring the integration of active learning with various model architectures (including deep neural networks, Gaussian processes, and language models), and addressing challenges like privacy preservation and handling open-set noise. This approach holds significant promise for reducing the substantial cost and effort of data labeling in diverse fields, ranging from image classification and natural language processing to materials science and healthcare.
Papers
Federated Active Learning for Target Domain Generalisation
Razvan Caramalau, Binod Bhattarai, Danail Stoyanov
Few Clicks Suffice: Active Test-Time Adaptation for Semantic Segmentation
Longhui Yuan, Shuang Li, Zhuo He, Binhui Xie
ActiveClean: Generating Line-Level Vulnerability Data via Active Learning
Ashwin Kallingal Joshy, Mirza Sanjida Alam, Shaila Sharmin, Qi Li, Wei Le
Comprehensive Benchmarking of Entropy and Margin Based Scoring Metrics for Data Selection
Anusha Sabbineni, Nikhil Anand, Maria Minakova
FreeAL: Towards Human-Free Active Learning in the Era of Large Language Models
Ruixuan Xiao, Yiwen Dong, Junbo Zhao, Runze Wu, Minmin Lin, Gang Chen, Haobo Wang
Active Foundational Models for Fault Diagnosis of Electrical Motors
Sriram Anbalagan, Sai Shashank GP, Deepesh Agarwal, Balasubramaniam Natarajan, Babji Srinivasan
FOCAL: A Cost-Aware Video Dataset for Active Learning
Kiran Kokilepersaud, Yash-Yee Logan, Ryan Benkert, Chen Zhou, Mohit Prabhushankar, Ghassan AlRegib, Enrique Corona, Kunjan Singh, Mostafa Parchami
RONAALP: Reduced-Order Nonlinear Approximation with Active Learning Procedure
Clément Scherding, Georgios Rigas, Denis Sipp, Peter J Schmid, Taraneh Sayadi