Sparse Training
Sparse training aims to reduce the computational cost and memory footprint of deep neural networks by training models with significantly fewer parameters, while maintaining or even improving accuracy. Current research focuses on developing efficient algorithms for creating and training sparse models, including methods for dynamic sparsity adjustment, improved initialization techniques, and hardware-accelerated computations, often applied to transformer and convolutional neural networks. These advancements are significant because they enable the deployment of large-scale models on resource-constrained devices and reduce the environmental impact of training, impacting both scientific research and practical applications in various fields.
Papers
May 31, 2023
May 28, 2023
May 3, 2023
April 27, 2023
April 15, 2023
April 14, 2023
February 18, 2023
February 9, 2023
February 6, 2023
January 9, 2023
December 2, 2022
November 30, 2022
November 29, 2022
November 26, 2022
November 14, 2022
September 22, 2022
September 20, 2022
September 11, 2022
August 11, 2022