Sparse Training
Sparse training aims to reduce the computational cost and memory footprint of deep neural networks by training models with significantly fewer parameters, while maintaining or even improving accuracy. Current research focuses on developing efficient algorithms for creating and training sparse models, including methods for dynamic sparsity adjustment, improved initialization techniques, and hardware-accelerated computations, often applied to transformer and convolutional neural networks. These advancements are significant because they enable the deployment of large-scale models on resource-constrained devices and reduce the environmental impact of training, impacting both scientific research and practical applications in various fields.
Papers
November 2, 2024
October 4, 2024
September 13, 2024
September 11, 2024
June 4, 2024
June 3, 2024
May 31, 2024
May 2, 2024
April 2, 2024
March 29, 2024
January 12, 2024
December 21, 2023
December 5, 2023
November 27, 2023
November 3, 2023
September 22, 2023
September 12, 2023
August 3, 2023