Contrastive Language Image
Contrastive Language-Image Pre-training (CLIP) models aim to learn joint representations of images and text, enabling zero-shot image classification and other multimodal tasks. Current research focuses on improving CLIP's localization capabilities, robustness to various data variations (including 3D data and low-light conditions), and efficiency through techniques like knowledge distillation and mixture-of-experts architectures. These advancements are significant for enhancing the reliability and applicability of CLIP in diverse fields, including medical image analysis, robotics, and AI-generated content detection.
Papers
February 26, 2024
February 23, 2024
February 15, 2024
February 12, 2024
February 9, 2024
February 6, 2024
January 24, 2024
January 15, 2024
January 12, 2024
January 4, 2024
January 3, 2024
January 2, 2024
December 22, 2023
December 20, 2023
December 19, 2023
December 18, 2023