Latent Diffusion Model
Latent diffusion models (LDMs) are generative AI models that create high-quality images by reversing a diffusion process in a compressed latent space, offering efficiency advantages over pixel-space methods. Current research focuses on improving controllability (e.g., through text or other modalities), enhancing efficiency (e.g., via parameter-efficient architectures or faster inference), and addressing challenges like model robustness and ethical concerns (e.g., watermarking and mitigating adversarial attacks). LDMs are significantly impacting various fields, including medical imaging (synthesis and restoration), speech enhancement, and even physics simulation, by enabling the generation of realistic and diverse data for training and analysis where real data is scarce or difficult to obtain.
Papers
SYRAC: Synthesize, Rank, and Count
Adriano D'Alessandro, Ali Mahdavi-Amiri, Ghassan Hamarneh
Toward effective protection against diffusion based mimicry through score distillation
Haotian Xue, Chumeng Liang, Xiaoyu Wu, Yongxin Chen
Sequential Data Generation with Groupwise Diffusion Process
Sangyun Lee, Gayoung Lee, Hyunsu Kim, Junho Kim, Youngjung Uh
Prompt-tuning latent diffusion models for inverse problems
Hyungjin Chung, Jong Chul Ye, Peyman Milanfar, Mauricio Delbracio
MaskDiffusion: Boosting Text-to-Image Consistency with Conditional Mask
Yupeng Zhou, Daquan Zhou, Zuo-Liang Zhu, Yaxing Wang, Qibin Hou, Jiashi Feng
Decoding visual brain representations from electroencephalography through Knowledge Distillation and latent diffusion models
Matteo Ferrante, Tommaso Boccato, Stefano Bargione, Nicola Toschi