Contrastive Language Image
Contrastive Language-Image Pre-training (CLIP) models aim to learn joint representations of images and text, enabling zero-shot image classification and other multimodal tasks. Current research focuses on improving CLIP's localization capabilities, robustness to various data variations (including 3D data and low-light conditions), and efficiency through techniques like knowledge distillation and mixture-of-experts architectures. These advancements are significant for enhancing the reliability and applicability of CLIP in diverse fields, including medical image analysis, robotics, and AI-generated content detection.
Papers
January 13, 2025
January 8, 2025
December 31, 2024
December 27, 2024
December 20, 2024
December 18, 2024
December 17, 2024
December 16, 2024
December 11, 2024
December 5, 2024
November 30, 2024
November 26, 2024
November 25, 2024
November 22, 2024
November 21, 2024
November 18, 2024
November 14, 2024
November 12, 2024
November 4, 2024