Pre Trained Text to Image
Pre-trained text-to-image models are revolutionizing image generation by leveraging vast datasets to create realistic and diverse images from textual descriptions. Current research focuses on enhancing controllability, addressing issues like artifact removal and multi-concept composition through techniques such as fine-tuning, adapter modules, and optimization of diffusion model parameters. This field is significant because it enables efficient image generation for various applications, from personalized image editing to creating high-quality videos and 3D models, while also pushing the boundaries of generative AI research.
Papers
Break-for-Make: Modular Low-Rank Adaptations for Composable Content-Style Customization
Yu Xu, Fan Tang, Juan Cao, Yuxin Zhang, Oliver Deussen, Weiming Dong, Jintao Li, Tong-Yee Lee
DreamSalon: A Staged Diffusion Framework for Preserving Identity-Context in Editable Face Generation
Haonan Lin, Mengmeng Wang, Yan Chen, Wenbin An, Yuzhe Yao, Guang Dai, Qianying Wang, Yong Liu, Jingdong Wang
An Optimization Framework to Enforce Multi-View Consistency for Texturing 3D Meshes
Zhengyi Zhao, Chen Song, Xiaodong Gu, Yuan Dong, Qi Zuo, Weihao Yuan, Liefeng Bo, Zilong Dong, Qixing Huang
DreamFlow: High-Quality Text-to-3D Generation by Approximating Probability Flow
Kyungmin Lee, Kihyuk Sohn, Jinwoo Shin