Input Compression
Input compression aims to reduce the computational burden of machine learning models by decreasing the size of input data without significant performance loss. Current research focuses on developing efficient compression techniques, including methods based on quasi-Monte Carlo point sets, dynamic networks with early exiting and quantization, and data augmentation strategies like positional consistency for transformers. These advancements are crucial for deploying large models on resource-constrained devices and improving training efficiency, impacting fields ranging from edge computing to large-scale model training.
Papers
September 20, 2024
June 20, 2024
November 22, 2023
November 8, 2022
August 5, 2022
July 19, 2022