Style PROMPT
Style Prompting research focuses on improving the performance and efficiency of large language models (LLMs) and other AI models by carefully crafting input prompts. Current research explores techniques like prompt tuning (adapting model behavior without full retraining), prompt baking (integrating prompts into model weights), and the use of multi-representation prompts to enhance model understanding and generalization across diverse tasks and modalities (e.g., vision-language models). This field is significant because effective prompting can drastically improve model performance, reduce computational costs, and mitigate biases, leading to more robust and reliable AI systems across various applications, including image processing, text generation, and even urban planning simulations.
Papers
Pedestrian Attribute Recognition via CLIP based Prompt Vision-Language Fusion
Xiao Wang, Jiandong Jin, Chenglong Li, Jin Tang, Cheng Zhang, Wei Wang
MM-TTS: Multi-modal Prompt based Style Transfer for Expressive Text-to-Speech Synthesis
Wenhao Guan, Yishuang Li, Tao Li, Hukai Huang, Feng Wang, Jiayan Lin, Lingyan Huang, Lin Li, Qingyang Hong