Active Learning
Active learning is a machine learning paradigm focused on optimizing data labeling efficiency by strategically selecting the most informative samples for annotation from a larger unlabeled pool. Current research emphasizes developing novel acquisition functions and data pruning strategies to reduce computational costs associated with large datasets, exploring the integration of active learning with various model architectures (including deep neural networks, Gaussian processes, and language models), and addressing challenges like privacy preservation and handling open-set noise. This approach holds significant promise for reducing the substantial cost and effort of data labeling in diverse fields, ranging from image classification and natural language processing to materials science and healthcare.
Papers
ActiveRAG: Autonomously Knowledge Assimilation and Accommodation through Retrieval-Augmented Agents
Zhipeng Xu, Zhenghao Liu, Yukun Yan, Shuo Wang, Shi Yu, Zheni Zeng, Chaojun Xiao, Zhiyuan Liu, Ge Yu, Chenyan Xiong
STENCIL: Submodular Mutual Information Based Weak Supervision for Cold-Start Active Learning
Nathan Beck, Adithya Iyer, Rishabh Iyer
Towards accelerating physical discovery via non-interactive and interactive multi-fidelity Bayesian Optimization: Current challenges and future opportunities
Arpan Biswas, Sai Mani Prudhvi Valleti, Rama Vasudevan, Maxim Ziatdinov, Sergei V. Kalinin
ELAD: Explanation-Guided Large Language Models Active Distillation
Yifei Zhang, Bo Pan, Chen Ling, Yuntong Hu, Liang Zhao
Mode Estimation with Partial Feedback
Charles Arnal, Vivien Cabannes, Vianney Perchet
Integrating Active Learning in Causal Inference with Interference: A Novel Approach in Online Experiments
Hongtao Zhu, Sizhe Zhang, Yang Su, Zhenyu Zhao, Nan Chen