Knowledge Distillation
Knowledge distillation is a machine learning technique that transfers knowledge from a large, complex "teacher" model to a smaller, more efficient "student" model, aiming to improve the student's performance and reduce computational costs. Current research focuses on improving distillation methods for various model architectures, including convolutional neural networks, transformers, and large language models, often incorporating techniques like parameter-efficient fine-tuning, multi-task learning, and data augmentation to enhance knowledge transfer. This approach is significant because it enables the deployment of high-performing models on resource-constrained devices and addresses challenges related to model size, training time, and privacy in diverse applications such as image captioning, speech processing, and medical diagnosis.
Papers - Page 49
Letz Translate: Low-Resource Machine Translation for Luxembourgish
Yewei Song, Saad Ezzini, Jacques Klein, Tegawende Bissyande, Clément Lefebvre, Anne GoujonDistillation from Heterogeneous Models for Top-K Recommendation
SeongKu Kang, Wonbin Kweon, Dongha Lee, Jianxun Lian, Xing Xie, Hwanjo YuDistilling Multi-Level X-vector Knowledge for Small-footprint Speaker Verification
Xuechen Liu, Md Sahidullah, Tomi Kinnunen
Towards domain generalisation in ASR with elitist sampling and ensemble knowledge distillation
Rehan Ahmad, Md Asif Jalal, Muhammad Umar Farooq, Anna Ollerenshaw, Thomas HainDistilled Reverse Attention Network for Open-world Compositional Zero-Shot Learning
Yun Li, Zhe Liu, Saurav Jha, Sally Cripps, Lina Yao
Practical Knowledge Distillation: Using DNNs to Beat DNNs
Chung-Wei Lee, Pavlos Athanasios Apostolopulos, Igor L. MarkovA Neural Span-Based Continual Named Entity Recognition Model
Yunan Zhang, Qingcai ChenPersonalized Decentralized Federated Learning with Knowledge Distillation
Eunjeong Jeong, Marios Kountouris