Text to Image Diffusion Model
Text-to-image diffusion models generate images from textual descriptions, aiming for high-fidelity and precise alignment. Current research focuses on improving controllability, addressing safety concerns (e.g., preventing generation of inappropriate content), and enhancing personalization capabilities through techniques like continual learning and latent space manipulation. These advancements are significant for various applications, including medical imaging, artistic creation, and data augmentation, while also raising important ethical considerations regarding model safety and bias.
Papers
Improving dermatology classifiers across populations using images generated by large diffusion models
Luke W. Sagers, James A. Diao, Matthew Groh, Pranav Rajpurkar, Adewole S. Adamson, Arjun K. Manrai
Peekaboo: Text to Image Diffusion Models are Zero-Shot Segmentors
Ryan Burgert, Kanchana Ranasinghe, Xiang Li, Michael S. Ryoo
Schr\"{o}dinger's Bat: Diffusion Models Sometimes Generate Polysemous Words in Superposition
Jennifer C. White, Ryan Cotterell