Latent Diffusion
Latent diffusion models are a class of generative models that synthesize data by reversing a diffusion process in a lower-dimensional latent space, offering advantages in efficiency and sample quality compared to direct data-space generation. Current research focuses on applying these models to diverse domains, including image generation, audio synthesis, medical image analysis, and scientific data modeling, often incorporating techniques like variational autoencoders and reinforcement learning for improved control and targeted generation. This approach is proving impactful by enabling data augmentation in data-scarce scenarios, facilitating the creation of high-fidelity synthetic data for training and evaluation, and offering new tools for analysis and interpretation in various scientific fields.
Papers
Kandinsky: an Improved Text-to-Image Synthesis with Image Prior and Latent Diffusion
Anton Razzhigaev, Arseniy Shakhmatov, Anastasia Maltseva, Vladimir Arkhipkin, Igor Pavlov, Ilya Ryabov, Angelina Kuts, Alexander Panchenko, Andrey Kuznetsov, Denis Dimitrov
Denoising Diffusion Step-aware Models
Shuai Yang, Yukang Chen, Luozhou Wang, Shu Liu, Yingcong Chen