Compression Task
Compression tasks in machine learning aim to reduce the size and computational cost of models while preserving performance. Current research focuses on developing efficient compression frameworks for various model architectures, including transformers and state-space models, often employing techniques like knowledge distillation, pruning, and reduced-order modeling. These advancements are crucial for deploying complex models on resource-constrained devices and improving the efficiency of large-scale applications such as image and video processing, natural language processing, and federated learning. The ultimate goal is to achieve a favorable trade-off between model size, computational speed, and accuracy.
Papers
November 15, 2024
September 26, 2024
August 6, 2024
July 15, 2024
July 5, 2024
May 24, 2024
May 11, 2024
April 15, 2024
March 14, 2024
March 10, 2024
February 23, 2024
December 12, 2023
November 10, 2023
August 26, 2022
July 5, 2022
June 2, 2022
April 14, 2022
March 30, 2022
January 31, 2022