Automatic Prompt Optimization

Automatic prompt optimization (APO) aims to automate the process of crafting effective prompts for large language models (LLMs), improving their performance on various tasks without requiring model retraining. Current research focuses on developing algorithms that iteratively refine prompts based on feedback, utilizing techniques like gradient-based optimization, reinforcement learning, and LLM-based self-reflection, often incorporating user-defined evaluation criteria or human feedback. APO's significance lies in its potential to significantly reduce the time and expertise needed for LLM deployment across diverse applications, from text summarization and image generation to clinical note creation and code search.

Papers