Efficient Active Learning
Efficient active learning aims to minimize the cost of labeling data for training machine learning models by strategically selecting the most informative unlabeled samples for annotation. Current research focuses on improving data selection strategies, often incorporating techniques like uncertainty sampling, diversity maximization, and leveraging pre-trained models (e.g., vision transformers, large language models) to accelerate the process. These advancements are crucial for addressing the high annotation costs associated with many machine learning tasks, particularly in domains like medical image analysis and autonomous driving, leading to more efficient and cost-effective model development.
Papers
Agnostic Active Learning of Single Index Models with Linear Sample Complexity
Aarshvi Gajjar, Wai Ming Tai, Xingyu Xu, Chinmay Hegde, Yi Li, Christopher Musco
Perception Without Vision for Trajectory Prediction: Ego Vehicle Dynamics as Scene Representation for Efficient Active Learning in Autonomous Driving
Ross Greer, Mohan Trivedi