Human Annotation
Human annotation, the process of labeling data for machine learning, is crucial but expensive and time-consuming. Current research focuses on mitigating this bottleneck through techniques like active learning, which prioritizes the most informative data points for human labeling, and the integration of large language models (LLMs) to automate or assist in the annotation process, including generating synthetic data or pre-annotating samples. These advancements aim to improve the efficiency and scalability of data annotation, ultimately accelerating the development and deployment of AI models across various domains, from natural language processing to medical image analysis. The resulting improvements in data quality and reduced annotation costs have significant implications for the broader AI research community and numerous practical applications.
Papers
Constructing Open Cloze Tests Using Generation and Discrimination Capabilities of Transformers
Mariano Felice, Shiva Taslimipoor, Paula Buttery
Scalable and Robust Self-Learning for Skill Routing in Large-Scale Conversational AI Systems
Mohammad Kachuee, Jinseok Nam, Sarthak Ahuja, Jin-Myung Won, Sungjin Lee