Attention Distillation
Attention distillation is a machine learning technique focused on transferring knowledge from a larger, more complex "teacher" model to a smaller, more efficient "student" model by focusing on the attention mechanisms within the teacher. Current research emphasizes applications across diverse areas, including image classification, object detection, semantic segmentation, and even graph neural networks, often employing vision transformers (ViTs) and convolutional neural networks (CNNs). This approach is significant for model compression, improving performance on resource-constrained devices, enhancing robustness (e.g., against backdoor attacks), and addressing challenges in continual learning and anomaly detection.
Papers
July 29, 2024
July 19, 2024
June 7, 2024
May 10, 2024
April 18, 2024
March 10, 2024
March 8, 2024
February 19, 2024
December 14, 2023
September 26, 2023
June 2, 2023
April 4, 2023
March 1, 2023
October 16, 2022
October 3, 2022
April 21, 2022
February 1, 2022