Zero Shot
Zero-shot learning aims to enable models to perform tasks on unseen data without any task-specific training, leveraging pre-trained knowledge to generalize to new situations. Current research focuses on improving zero-shot capabilities across diverse modalities (vision, language, audio) using large language models (LLMs), vision-language models (VLMs), and diffusion models, often incorporating techniques like chain-of-thought prompting, knowledge retrieval, and prompt engineering to enhance performance and interpretability. This field is significant because it promises more efficient and adaptable AI systems, impacting various applications from image editing and medical diagnosis to robotics and natural language processing.
Papers
MotionShop: Zero-Shot Motion Transfer in Video Diffusion Models with Mixture of Score Guidance
Hidir Yesiltepe, Tuna Han Salih Meral, Connor Dunlop, Pinar Yanardag
Prompt Transfer for Dual-Aspect Cross Domain Cognitive Diagnosis
Fei Liu, Yizhong Zhang, Shuochen Liu, Shengwei Ji, Kui Yu, Le Wu
$S^3$: Synonymous Semantic Space for Improving Zero-Shot Generalization of Vision-Language Models
Xiaojie Yin, Qilong Wang, Bing Cao, Qinghua Hu
Grounding Descriptions in Images informs Zero-Shot Visual Recognition
Shaunak Halbe, Junjiao Tian, K J Joseph, James Seale Smith, Katherine Stevo, Vineeth N Balasubramanian, Zsolt Kira
SeeGround: See and Ground for Zero-Shot Open-Vocabulary 3D Visual Grounding
Rong Li, Shijie Li, Lingdong Kong, Xulei Yang, Junwei Liang
DynRank: Improving Passage Retrieval with Dynamic Zero-Shot Prompting Based on Question Classification
Abdelrahman Abdallah, Jamshid Mozafari, Bhawna Piryani, Mohammed M.Abdelgwad, Adam Jatowt
Enhancing Zero-shot Chain of Thought Prompting via Uncertainty-Guided Strategy Selection
Shanu Kumar, Saish Mendke, Karody Lubna Abdul Rahman, Santosh Kurasa, Parag Agrawal, Sandipan Dandapat
T2Vid: Translating Long Text into Multi-Image is the Catalyst for Video-LLMs
Shukang Yin, Chaoyou Fu, Sirui Zhao, Yunhang Shen, Chunjiang Ge, Yan Yang, Zuwei Long, Yuhan Dai, Tong Xu, Xing Sun, Ran He, Caifeng Shan, Enhong Chen
Dual Risk Minimization: Towards Next-Level Robustness in Fine-tuning Zero-Shot Models
Kaican Li, Weiyan Xie, Yongxiang Huang, Didan Deng, Lanqing Hong, Zhenguo Li, Ricardo Silva, Nevin L. Zhang
Diorama: Unleashing Zero-shot Single-view 3D Scene Modeling
Qirui Wu, Denys Iliash, Daniel Ritchie, Manolis Savva, Angel X. Chang
Proto Successor Measure: Representing the Space of All Possible Solutions of Reinforcement Learning
Siddhant Agarwal, Harshit Sikchi, Peter Stone, Amy Zhang
Zero-Forget Preservation of Semantic Communication Alignment in Distributed AI Networks
Jingzhi Hu, Geoffrey Ye Li
Automatic Prompt Generation and Grounding Object Detection for Zero-Shot Image Anomaly Detection
Tsun-Hin Cheung, Ka-Chun Fung, Songjiang Lai, Kwan-Ho Lin, Vincent Ng, Kin-Man Lam
Track Anything Behind Everything: Zero-Shot Amodal Video Object Segmentation
Finlay G. C. Hudson, William A. P. Smith
Relation-Aware Meta-Learning for Zero-shot Sketch-Based Image Retrieval
Yang Liu, Jiale Du, Xinbo Gao, Jungong Han
CoDiff-VC: A Codec-Assisted Diffusion Model for Zero-shot Voice Conversion
Yuke Li, Xinfa Zhu, Hanzhao Li, JiXun Yao, WenJie Tian, YunLin Chen, YunLin Chen, Zhifei Li, Lei Xie