Learnable Prompt
Learnable prompts are adaptable input sequences used to guide large pre-trained models, such as vision-language models (VLMs) and Segment Anything Models (SAMs), towards specific tasks without extensive model retraining. Current research focuses on improving the robustness and generalization of these prompts across diverse datasets and tasks, often employing techniques like prompt refinement, hybrid prompt architectures (static and dynamic), and knowledge distillation. This approach offers a parameter-efficient way to adapt powerful foundation models to various applications, including medical image analysis, anomaly detection, and few-shot learning, thereby enhancing their practicality and reducing the need for large labeled datasets.
Papers
CPL: Counterfactual Prompt Learning for Vision and Language Models
Xuehai He, Diji Yang, Weixi Feng, Tsu-Jui Fu, Arjun Akula, Varun Jampani, Pradyumna Narayana, Sugato Basu, William Yang Wang, Xin Eric Wang
Continued Pretraining for Better Zero- and Few-Shot Promptability
Zhaofeng Wu, Robert L. Logan, Pete Walsh, Akshita Bhagia, Dirk Groeneveld, Sameer Singh, Iz Beltagy