Contrastive Language Image
Contrastive Language-Image Pre-training (CLIP) models aim to learn joint representations of images and text, enabling zero-shot image classification and other multimodal tasks. Current research focuses on improving CLIP's localization capabilities, robustness to various data variations (including 3D data and low-light conditions), and efficiency through techniques like knowledge distillation and mixture-of-experts architectures. These advancements are significant for enhancing the reliability and applicability of CLIP in diverse fields, including medical image analysis, robotics, and AI-generated content detection.
Papers
November 4, 2024
November 1, 2024
October 30, 2024
October 16, 2024
October 12, 2024
October 11, 2024
October 8, 2024
October 3, 2024
October 2, 2024
October 1, 2024
September 28, 2024
September 26, 2024
September 23, 2024
September 20, 2024
September 19, 2024
September 15, 2024
September 10, 2024