Zero Shot
Zero-shot learning aims to enable models to perform tasks on unseen data without any task-specific training, leveraging pre-trained knowledge to generalize to new situations. Current research focuses on improving zero-shot capabilities across diverse modalities (vision, language, audio) using large language models (LLMs), vision-language models (VLMs), and diffusion models, often incorporating techniques like chain-of-thought prompting, knowledge retrieval, and prompt engineering to enhance performance and interpretability. This field is significant because it promises more efficient and adaptable AI systems, impacting various applications from image editing and medical diagnosis to robotics and natural language processing.
1164papers
Papers - Page 29
June 5, 2024
Mind's Eye: Image Recognition by EEG via Multimodal Similarity-Keeping Contrastive Learning
Balancing Performance and Efficiency in Zero-shot Robotic Navigation
ZeroDiff: Solidified Visual-Semantic Correlation in Zero-Shot Learning
Visual-Text Cross Alignment: Refining the Similarity Score in Vision-Language Models
DiffCut: Catalyzing Zero-Shot Semantic Segmentation with Diffusion Features and Recursive Normalized Cut
June 4, 2024
June 3, 2024
June 1, 2024
May 30, 2024