Image Generation Model
Image generation models aim to create realistic and diverse images from various inputs like text, sketches, or other images. Current research focuses on improving model architectures (e.g., diffusion models, transformers, hybrid approaches) to enhance image quality, controllability, and efficiency, while also addressing challenges like bias mitigation, prompt engineering, and the generation of specific image types (e.g., RGBA images, 3D models). These advancements have significant implications for various fields, including accessible communication, creative design, and data augmentation, but also raise important ethical considerations regarding bias and safety. The ongoing development of robust and responsible image generation models is a key area of active research.
Papers
FlexiFilm: Long Video Generation with Flexible Conditions
Yichen Ouyang, jianhao Yuan, Hao Zhao, Gaoang Wang, Bo zhao
Anywhere: A Multi-Agent Framework for Reliable and Diverse Foreground-Conditioned Image Inpainting
Tianyidan Xie, Rui Ma, Qian Wang, Xiaoqian Ye, Feixuan Liu, Ying Tai, Zhenyu Zhang, Zili Yi
Understanding Subjectivity through the Lens of Motivational Context in Model-Generated Image Satisfaction
Senjuti Dutta, Sherol Chen, Sunny Mak, Amnah Ahmad, Katherine Collins, Alena Butryna, Deepak Ramachandran, Krishnamurthy Dvijotham, Ellie Pavlick, Ravi Rajakumar
T-HITL Effectively Addresses Problematic Associations in Image Generation and Maintains Overall Visual Quality
Susan Epstein, Li Chen, Alessandro Vecchiato, Ankit Jain