Data Free Knowledge Distillation

Data-free knowledge distillation (DFKD) focuses on transferring knowledge from a large, pre-trained "teacher" model to a smaller "student" model without access to the original training data. Current research emphasizes developing robust methods to generate high-quality synthetic data for student training, often employing generative adversarial networks (GANs), diffusion models, or other generative techniques, and addressing issues like distribution shifts between synthetic and real data. This field is significant because it enables model compression, enhances privacy, and facilitates knowledge transfer in situations where original data is unavailable or restricted, impacting various applications from medical image analysis to federated learning.

Papers