Prompt Diffusion

Prompt diffusion leverages diffusion models to generate images and other data modalities (audio, text) conditioned on various inputs, primarily text prompts but increasingly incorporating visual context and other modalities. Current research focuses on improving in-context learning within these models, enhancing controllability through prompt engineering and embedding manipulation, and exploring prompt-free approaches using only visual input. This rapidly evolving field is significantly impacting image generation, semantic segmentation, and other areas by enabling more flexible, controllable, and efficient generation of high-quality data, particularly in scenarios with limited labeled data.

Papers