Zero Shot
Zero-shot learning aims to enable models to perform tasks on unseen data without any task-specific training, leveraging pre-trained knowledge to generalize to new situations. Current research focuses on improving zero-shot capabilities across diverse modalities (vision, language, audio) using large language models (LLMs), vision-language models (VLMs), and diffusion models, often incorporating techniques like chain-of-thought prompting, knowledge retrieval, and prompt engineering to enhance performance and interpretability. This field is significant because it promises more efficient and adaptable AI systems, impacting various applications from image editing and medical diagnosis to robotics and natural language processing.
Papers
VLN-Game: Vision-Language Equilibrium Search for Zero-Shot Semantic Navigation
Bangguo Yu, Yuzhen Liu, Lei Han, Hamidreza Kasaei, Tingguang Li, Ming Cao
Zero-Shot Load Forecasting with Large Language Models
Wenlong Liao, Zhe Yang, Mengshuo Jia, Christian Rehtanz, Jiannong Fang, Fernando Porté-Agel
Neuron: Learning Context-Aware Evolving Representations for Zero-Shot Skeleton Action Recognition
Yang Chen, Jingcai Guo, Song Guo, Dacheng Tao
Zero-Shot Automatic Annotation and Instance Segmentation using LLM-Generated Datasets: Eliminating Field Imaging and Manual Annotation for Deep Learning Model Development
Ranjan Sapkota, Achyut Paudel, Manoj Karkee
ZeFaV: Boosting Large Language Models for Zero-shot Fact Verification
Son T. Luu, Hiep Nguyen, Trung Vo, Le-Minh Nguyen
Retrieval Augmented Time Series Forecasting
Kutay Tire, Ege Onur Taga, Muhammed Emrullah Ildiz, Samet Oymak
Latent Space Disentanglement in Diffusion Transformers Enables Precise Zero-shot Semantic Editing
Zitao Shuai, Chenwei Wu, Zhengxu Tang, Bowen Song, Liyue Shen
LLMPhy: Complex Physical Reasoning Using Large Language Models and World Models
Anoop Cherian, Radu Corcodel, Siddarth Jain, Diego Romeres
Audiobox TTA-RAG: Improving Zero-Shot and Few-Shot Text-To-Audio with Retrieval-Augmented Generation
Mu Yang, Bowen Shi, Matthew Le, Wei-Ning Hsu, Andros Tjandra
SG-I2V: Self-Guided Trajectory Control in Image-to-Video Generation
Koichi Namekata, Sherwin Bahmani, Ziyi Wu, Yash Kant, Igor Gilitschenski, David B. Lindell
DINO-WM: World Models on Pre-trained Visual Features enable Zero-shot Planning
Gaoyue Zhou, Hengkai Pan, Yann LeCun, Lerrel Pinto
Zero-Shot Temporal Resolution Domain Adaptation for Spiking Neural Networks
Sanja Karilanova, Maxime Fabre, Emre Neftci, Ayça Özçelikkale
Best Practices for Distilling Large Language Models into BERT for Web Search Ranking
Dezhi Ye, Junwei Hu, Jiabin Fan, Bowen Tian, Jie Liu, Haijin Liang, Jin Ma