Low Temperature Distillation
Low-temperature distillation, a model compression technique, aims to transfer knowledge from a large, computationally expensive "teacher" model to a smaller, more efficient "student" model. Current research focuses on improving distillation methods across diverse architectures (including CNNs, Transformers, and GNNs), addressing challenges like mitigating backdoors in teacher models, handling data scarcity, and achieving robust performance across different datasets and tasks. These advancements are significant for deploying complex models on resource-constrained devices and improving the efficiency and scalability of machine learning applications.
Papers
Latent Dataset Distillation with Diffusion Models
Brian B. Moser, Federico Raue, Sebastian Palacio, Stanislav Frolov, Andreas Dengel
On the Effectiveness of Distillation in Mitigating Backdoors in Pre-trained Encoder
Tingxu Han, Shenghan Huang, Ziqi Ding, Weisong Sun, Yebo Feng, Chunrong Fang, Jun Li, Hanwei Qian, Cong Wu, Quanjun Zhang, Yang Liu, Zhenyu Chen