Input Compression

Input compression aims to reduce the computational burden of machine learning models by decreasing the size of input data without significant performance loss. Current research focuses on developing efficient compression techniques, including methods based on quasi-Monte Carlo point sets, dynamic networks with early exiting and quantization, and data augmentation strategies like positional consistency for transformers. These advancements are crucial for deploying large models on resource-constrained devices and improving training efficiency, impacting fields ranging from edge computing to large-scale model training.

Papers