Diffusion Based Data Augmentation
Diffusion-based data augmentation leverages the power of generative diffusion models to create synthetic training data, addressing the limitations of small or imbalanced datasets in various machine learning tasks. Current research focuses on improving the fidelity and diversity of generated data, often incorporating techniques like conditional generation with text or image prompts, and integrating diffusion models with other architectures such as VAEs or LLMs to enhance control and semantic consistency. This approach holds significant promise for improving the performance and robustness of machine learning models across diverse applications, particularly in domains with limited labeled data, such as medical imaging and object detection.
Papers
DifAugGAN: A Practical Diffusion-style Data Augmentation for GAN-based Single Image Super-resolution
Axi Niu, Kang Zhang, Joshua Tian Jin Tee, Trung X. Pham, Jinqiu Sun, Chang D. Yoo, In So Kweon, Yanning Zhang
Prompt-Based Exemplar Super-Compression and Regeneration for Class-Incremental Learning
Ruxiao Duan, Yaoyao Liu, Jieneng Chen, Adam Kortylewski, Alan Yuille