Image Synthesis
Image synthesis focuses on generating realistic images from various inputs, such as text descriptions, sketches, or other images, aiming to improve controllability, realism, and efficiency. Current research emphasizes advancements in diffusion models, generative adversarial networks (GANs), and autoregressive models, often incorporating techniques like latent space manipulation, multimodal conditioning (text and image), and attention mechanisms to enhance image quality and control. This field is significant for its applications in diverse areas, including medical imaging, virtual try-ons, and content creation, while also raising important considerations regarding ethical implications and environmental impact of computationally intensive models.
Papers
Affordance Diffusion: Synthesizing Hand-Object Interactions
Yufei Ye, Xueting Li, Abhinav Gupta, Shalini De Mello, Stan Birchfield, Jiaming Song, Shubham Tulsiani, Sifei Liu
DiffuMask: Synthesizing Images with Pixel-level Annotations for Semantic Segmentation Using Diffusion Models
Weijia Wu, Yuzhong Zhao, Mike Zheng Shou, Hong Zhou, Chunhua Shen
P+: Extended Textual Conditioning in Text-to-Image Generation
Andrey Voynov, Qinghao Chu, Daniel Cohen-Or, Kfir Aberman
DiffIR: Efficient Diffusion Model for Image Restoration
Bin Xia, Yulun Zhang, Shiyin Wang, Yitong Wang, Xinglong Wu, Yapeng Tian, Wenming Yang, Luc Van Gool
SpectralCLIP: Preventing Artifacts in Text-Guided Style Transfer from a Spectral Perspective
Zipeng Xu, Songlong Xing, Enver Sangineto, Nicu Sebe
Painting 3D Nature in 2D: View Synthesis of Natural Scenes from a Single Semantic Mask
Shangzhan Zhang, Sida Peng, Tianrun Chen, Linzhan Mou, Haotong Lin, Kaicheng Yu, Yiyi Liao, Xiaowei Zhou
Text-Guided Scene Sketch-to-Photo Synthesis
AprilPyone MaungMaung, Makoto Shing, Kentaro Mitsui, Kei Sawada, Fumio Okura