Contrastive Language Image
Contrastive Language-Image Pre-training (CLIP) models aim to learn joint representations of images and text, enabling zero-shot image classification and other multimodal tasks. Current research focuses on improving CLIP's localization capabilities, robustness to various data variations (including 3D data and low-light conditions), and efficiency through techniques like knowledge distillation and mixture-of-experts architectures. These advancements are significant for enhancing the reliability and applicability of CLIP in diverse fields, including medical image analysis, robotics, and AI-generated content detection.
Papers
June 22, 2022
June 4, 2022
May 28, 2022
May 23, 2022
May 3, 2022
April 29, 2022
April 7, 2022
March 21, 2022
March 15, 2022
March 11, 2022
January 26, 2022
January 15, 2022
December 27, 2021
December 16, 2021
December 15, 2021
December 9, 2021
December 4, 2021
December 2, 2021