Layout to Image Synthesis
Layout-to-image synthesis aims to generate realistic images from user-provided layouts specifying object locations and categories, offering finer control than text-based methods. Current research focuses on training-free approaches using pre-trained diffusion models, often incorporating attention mechanisms to improve object placement accuracy and semantic coherence, and addressing challenges like handling unseen object classes and ensuring consistent object relationships. This field is significant for advancing image generation capabilities, enabling more intuitive and precise control over image content for applications ranging from creative design tools to virtual and augmented reality.
Papers
October 9, 2024
July 18, 2024
November 21, 2023
November 17, 2023
March 25, 2023
August 12, 2022
April 5, 2022