Latent Diffusion
Latent diffusion models are a class of generative models that synthesize data by reversing a diffusion process in a lower-dimensional latent space, offering advantages in efficiency and sample quality compared to direct data-space generation. Current research focuses on applying these models to diverse domains, including image generation, audio synthesis, medical image analysis, and scientific data modeling, often incorporating techniques like variational autoencoders and reinforcement learning for improved control and targeted generation. This approach is proving impactful by enabling data augmentation in data-scarce scenarios, facilitating the creation of high-fidelity synthetic data for training and evaluation, and offering new tools for analysis and interpretation in various scientific fields.
Papers
Inpainting Pathology in Lumbar Spine MRI with Latent Diffusion
Colin Hansen, Simas Glinskis, Ashwin Raju, Micha Kornreich, JinHyeong Park, Jayashri Pawar, Richard Herzog, Li Zhang, Benjamin Odry
MoLA: Motion Generation and Editing with Latent Diffusion Enhanced by Adversarial Training
Kengo Uchida, Takashi Shibuya, Yuhta Takida, Naoki Murata, Shusuke Takahashi, Yuki Mitsufuji
Full Event Particle-Level Unfolding with Variable-Length Latent Variational Diffusion
Alexander Shmakov, Kevin Greif, Michael James Fenton, Aishik Ghosh, Pierre Baldi, Daniel Whiteson
FLDM-VTON: Faithful Latent Diffusion Model for Virtual Try-on
Chenhui Wang, Tao Chen, Zhihao Chen, Zhizhong Huang, Taoran Jiang, Qi Wang, Hongming Shan