Image Synthesis
Image synthesis focuses on generating realistic images from various inputs, such as text descriptions, sketches, or other images, aiming to improve controllability, realism, and efficiency. Current research emphasizes advancements in diffusion models, generative adversarial networks (GANs), and autoregressive models, often incorporating techniques like latent space manipulation, multimodal conditioning (text and image), and attention mechanisms to enhance image quality and control. This field is significant for its applications in diverse areas, including medical imaging, virtual try-ons, and content creation, while also raising important considerations regarding ethical implications and environmental impact of computationally intensive models.
Papers
UNIC-Adapter: Unified Image-instruction Adapter with Multi-modal Transformer for Image Generation
Lunhao Duan, Shanshan Zhao, Wenjun Yan, Yinglun Li, Qing-Guo Chen, Zhao Xu, Weihua Luo, Kaifu Zhang, Mingming Gong, Gui-Song Xia
DRDM: A Disentangled Representations Diffusion Model for Synthesizing Realistic Person Images
Enbo Huang, Yuan Zhang, Faliang Huang, Guangyu Zhang, Yang Liu
Infinity: Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
Jian Han, Jinlai Liu, Yi Jiang, Bin Yan, Yuqi Zhang, Zehuan Yuan, Bingyue Peng, Xiaobing Liu
A Framework For Image Synthesis Using Supervised Contrastive Learning
Yibin Liu, Jianyu Zhang, Li Zhang, Shijian Li, Gang Pan
Reconciling Semantic Controllability and Diversity for Remote Sensing Image Synthesis with Hybrid Semantic Embedding
Junde Liu, Danpei Zhao, Bo Yuan, Wentao Li, Tian Li
Cross Group Attention and Group-wise Rolling for Multimodal Medical Image Synthesis
Tao Song, Yicheng Wu, Minhao Hu, Xiangde Luo, Linda Wei, Guotai Wang, Yi Guo, Feng Xu, Shaoting Zhang