Zero Shot
Zero-shot learning aims to enable models to perform tasks on unseen data without any task-specific training, leveraging pre-trained knowledge to generalize to new situations. Current research focuses on improving zero-shot capabilities across diverse modalities (vision, language, audio) using large language models (LLMs), vision-language models (VLMs), and diffusion models, often incorporating techniques like chain-of-thought prompting, knowledge retrieval, and prompt engineering to enhance performance and interpretability. This field is significant because it promises more efficient and adaptable AI systems, impacting various applications from image editing and medical diagnosis to robotics and natural language processing.
Papers
DeCap: Decoding CLIP Latents for Zero-Shot Captioning via Text-Only Training
Wei Li, Linchao Zhu, Longyin Wen, Yi Yang
Model Sketching: Centering Concepts in Early-Stage Machine Learning Model Design
Michelle S. Lam, Zixian Ma, Anne Li, Izequiel Freitas, Dakuo Wang, James A. Landay, Michael S. Bernstein
CHiLL: Zero-shot Custom Interpretable Feature Extraction from Clinical Notes with Large Language Models
Denis Jered McInerney, Geoffrey Young, Jan-Willem van de Meent, Byron C. Wallace
Region-Aware Diffusion for Zero-shot Text-driven Image Editing
Nisha Huang, Fan Tang, Weiming Dong, Tong-Yee Lee, Changsheng Xu
A Systematic Evaluation and Benchmark for Embedding-Aware Generative Models: Features, Models, and Any-shot Scenarios
Liangjun Feng, Jiancheng Zhao, Chunhui Zhao
Zero-shot Sim2Real Adaptation Across Environments
Buddhika Laknath Semage, Thommen George Karimpanal, Santu Rana, Svetha Venkatesh
Is ChatGPT a General-Purpose Natural Language Processing Task Solver?
Chengwei Qin, Aston Zhang, Zhuosheng Zhang, Jiaao Chen, Michihiro Yasunaga, Diyi Yang