Pixel Level Annotation
Pixel-level annotation, the process of labeling each pixel in an image with a specific class, is crucial for training many computer vision models, particularly those for semantic segmentation. Current research focuses on reducing the substantial manual effort required for this task, exploring techniques like weak supervision (using bounding boxes or points instead of full pixel masks), semi-supervised learning (combining labeled and unlabeled data), and active learning (iteratively selecting the most informative pixels to label). These advancements leverage various model architectures, including transformers, convolutional neural networks, and diffusion models, to improve annotation efficiency and accuracy. The resulting improvements in annotation efficiency have significant implications for various fields, enabling the development of more accurate and robust models for applications ranging from medical image analysis to autonomous driving.
Papers
Multi-Scale Attention-based Multiple Instance Learning for Classification of Multi-Gigapixel Histology Images
Made Satria Wibawa, Kwok-Wai Lo, Lawrence Young, Nasir Rajpoot
A Weakly Supervised Learning Framework for Salient Object Detection via Hybrid Labels
Runmin Cong, Qi Qin, Chen Zhang, Qiuping Jiang, Shiqi Wang, Yao Zhao, Sam Kwong