Adversarial Distillation

Adversarial distillation is a machine learning technique that enhances the robustness and accuracy of smaller "student" models by transferring knowledge from larger, more robust "teacher" models, often trained with adversarial examples. Current research focuses on improving knowledge transfer efficiency through methods like dynamic guidance, feature-level distillation, and adversarial training incorporating dynamic labels or gradient matching. This approach is significant for improving model performance in resource-constrained environments and for bolstering defenses against adversarial attacks, with applications ranging from image classification and natural language processing to medical image analysis and 3D generation.

Papers