High Sparsity
High sparsity in neural networks focuses on reducing the number of parameters while maintaining or improving model performance and efficiency. Current research explores various techniques, including structured pruning methods (e.g., block-sparse architectures and NxM sparsity), zeroth-order optimization for memory-constrained fine-tuning, and novel training algorithms that promote sparsity from scratch or during iterative pruning. This research is significant because it addresses the computational and memory limitations of large models, enabling deployment on resource-constrained devices and improving the energy efficiency of AI systems.
Papers
October 15, 2024
September 5, 2024
July 24, 2024
June 5, 2024
April 3, 2024
February 7, 2024
December 21, 2023
November 28, 2023
October 10, 2023
October 8, 2023
February 3, 2023
December 4, 2022
November 14, 2022
October 25, 2022
October 6, 2022
June 30, 2022
June 22, 2022
June 14, 2022