Model Sparsity
Model sparsity focuses on reducing the number of parameters or activations in machine learning models to improve efficiency and resource utilization without significant performance loss. Current research explores various techniques, including weight pruning, activation sparsity, and the incorporation of sparsity-inducing regularizations during training, applied to diverse architectures like large language models and convolutional neural networks. This area is crucial for deploying large models on resource-constrained devices and improving the interpretability and robustness of machine learning systems, impacting both scientific understanding and practical applications across various domains.
Papers
August 26, 2024
March 17, 2024
February 25, 2024
November 9, 2023
August 14, 2023
May 4, 2023
April 11, 2023
August 11, 2022
August 8, 2022
July 6, 2022
June 17, 2022