Compression Framework
Compression frameworks are being actively developed to address the growing demands of large-scale machine learning models and high-resolution data streams, aiming to reduce computational costs and memory usage without significant performance loss. Current research focuses on techniques like knowledge distillation, model pruning, and the development of novel compression algorithms tailored to specific data types (e.g., video, spike streams, embeddings) and model architectures (e.g., Transformers). These advancements are crucial for deploying complex models on resource-constrained devices and improving the efficiency of various applications, including recommendation systems, video processing, and distributed machine learning.
Papers
June 7, 2024
May 20, 2024
December 6, 2023
August 15, 2023
June 25, 2023
May 24, 2023
September 9, 2022
June 30, 2022