Network Compression
Network compression aims to reduce the size and computational cost of deep neural networks (DNNs) without significant performance loss. Current research focuses on techniques like pruning (removing less important connections), quantization (reducing the precision of weights), and low-rank approximations, often applied during training or post-training, and applied to various architectures including CNNs, GANs, and transformers. These advancements are crucial for deploying large-scale DNNs on resource-constrained devices and improving the efficiency of training and inference, impacting both scientific understanding of DNNs and their practical applications across diverse fields.
Papers
June 27, 2024
May 19, 2024
May 15, 2024
May 6, 2024
March 1, 2024
February 29, 2024
September 29, 2023
June 9, 2023
June 5, 2023
May 25, 2023
April 1, 2023
March 16, 2023
March 4, 2023
March 2, 2023
February 15, 2023
September 15, 2022
September 9, 2022
July 6, 2022
June 15, 2022