Image Generation
Image generation research focuses on creating realistic and diverse images from various inputs, such as text, sketches, or other images, aiming for greater control and efficiency. Current efforts center on refining diffusion and autoregressive models, exploring techniques like dynamic computation, disentangled feature representation, and multimodal integration to improve image quality, controllability, and computational efficiency. These advancements have significant implications for accessible communication, creative content production, and various computer vision tasks, offering powerful tools for both scientific investigation and practical applications. Ongoing work addresses challenges like handling multiple conditions, improving evaluation metrics, and mitigating biases and limitations in existing models.
Papers
UniFL: Improve Latent Diffusion Model via Unified Feedback Learning
Jiacheng Zhang, Jie Wu, Yuxi Ren, Xin Xia, Huafeng Kuang, Pan Xie, Jiashi Li, Xuefeng Xiao, Weilin Huang, Shilei Wen, Lean Fu, Guanbin Li
Mask-ControlNet: Higher-Quality Image Generation with An Additional Mask Prompt
Zhiqi Huang, Huixin Xiong, Haoyu Wang, Longguang Wang, Zhiheng Li
Reference-Based 3D-Aware Image Editing with Triplanes
Bahri Batuhan Bilecen, Yigit Yalin, Ning Yu, Aysegul Dundar
Robust Concept Erasure Using Task Vectors
Minh Pham, Kelly O. Marshall, Chinmay Hegde, Niv Cohen
Diverse and Tailored Image Generation for Zero-shot Multi-label Classification
Kaixin Zhang, Zhixiang Yuan, Tao Huang