Single CLIP
Single CLIP, a powerful vision-language model, is being extensively studied to improve its performance and address its limitations in various applications. Current research focuses on mitigating issues like object hallucinations, enhancing its capabilities for specialized domains (e.g., agriculture), and developing robust defenses against adversarial attacks and biases. This work is significant because it explores ways to leverage CLIP's impressive zero-shot capabilities while simultaneously improving its accuracy, reliability, and fairness across diverse downstream tasks, impacting fields ranging from image generation to anomaly detection.
Papers
May 8, 2024
May 4, 2024
May 2, 2024
May 1, 2024
April 19, 2024
April 18, 2024
April 16, 2024
April 9, 2024
April 8, 2024
April 4, 2024
April 2, 2024
March 31, 2024
March 30, 2024
March 27, 2024
March 25, 2024
March 22, 2024
March 15, 2024