Model Sparsification
Model sparsification aims to reduce the size and computational cost of machine learning models, particularly deep neural networks, without significant performance loss. Current research focuses on developing efficient algorithms for pruning model weights, often incorporating techniques like L0 regularization, Bayesian methods, and structured sparsity to achieve high sparsity levels in various architectures, including ResNets, Transformers, and other convolutional and graph neural networks. This work is driven by the need to deploy large models on resource-constrained devices and improve training efficiency in federated learning settings, impacting both the scalability of AI and its energy consumption.
Papers
September 5, 2024
September 2, 2024
August 30, 2024
February 25, 2024
January 12, 2024
December 20, 2023
December 3, 2023
October 30, 2023
August 23, 2023
April 26, 2023
May 8, 2022
April 26, 2022