Zero Shot Learning
Zero-shot learning (ZSL) aims to enable machine learning models to classify data from unseen categories without requiring any training examples for those categories, leveraging knowledge transferred from seen categories. Current research focuses on improving ZSL performance across various modalities (image, text, audio, graph data) using large language models (LLMs), vision-language models (VLMs), and graph neural networks (GNNs), often incorporating techniques like prompt engineering and contrastive learning. This capability is highly significant for addressing data scarcity issues in many fields, including medical image analysis, natural language processing, and robotics, enabling more efficient and adaptable AI systems. The development of more efficient and robust ZSL methods is a key area of ongoing research.
Papers
CICA: Content-Injected Contrastive Alignment for Zero-Shot Document Image Classification
Sankalp Sinha, Muhammad Saif Ullah Khan, Talha Uddin Sheikh, Didier Stricker, Muhammad Zeshan Afzal
Dual Relation Mining Network for Zero-Shot Learning
Jinwei Han, Yingguo Gao, Zhiwen Lin, Ke Yan, Shouhong Ding, Yuan Gao, Gui-Song Xia
Enhancing Q-Learning with Large Language Model Heuristics
Xiefeng Wu
The Devil is in the Few Shots: Iterative Visual Knowledge Completion for Few-shot Learning
Yaohui Li, Qifeng Zhou, Haoxing Chen, Jianbing Zhang, Xinyu Dai, Hao Zhou
CREST: Cross-modal Resonance through Evidential Deep Learning for Enhanced Zero-Shot Learning
Haojian Huang, Xiaozhen Qiao, Zhuo Chen, Haodong Chen, Bingyu Li, Zhe Sun, Mulin Chen, Xuelong Li