Compositional Image Generation
Compositional image generation focuses on creating images that accurately reflect the combination of multiple concepts described in text or other input modalities. Current research emphasizes improving the ability of models, often based on diffusion processes or discrete generative models, to handle complex compositions involving objects, attributes, and spatial relationships, often using large vision-language models for evaluation and refinement. This area is significant because it pushes the boundaries of AI's understanding of visual semantics and has implications for various applications, including advanced image editing, content creation, and more robust visual question answering systems.
Papers
November 12, 2024
August 26, 2024
May 10, 2024
October 10, 2023
September 25, 2023
June 1, 2023
April 25, 2023
April 5, 2023