Conditional Image Generation
Conditional image generation aims to synthesize images based on various conditions, such as text descriptions, sketches, or semantic maps, striving for high fidelity and realism. Current research focuses on improving control over the generation process through techniques like latent space manipulation, diffusion models (including variations like transformers and ODE-based approaches), and integrating diverse conditioning modalities (e.g., combining text with sketches or depth maps). These advancements are significant for applications ranging from creative content generation to image editing and enhancement, driving progress in both computer vision and generative modeling.
Papers
Diffusion 3D Features (Diff3F): Decorating Untextured Shapes with Distilled Semantic Features
Niladri Shekhar Dutt, Sanjeev Muralikrishnan, Niloy J. Mitra
Manifold Preserving Guided Diffusion
Yutong He, Naoki Murata, Chieh-Hsin Lai, Yuhta Takida, Toshimitsu Uesaka, Dongjun Kim, Wei-Hsiang Liao, Yuki Mitsufuji, J. Zico Kolter, Ruslan Salakhutdinov, Stefano Ermon