Unstructured Sparsity
Unstructured sparsity focuses on making neural networks more efficient by removing individual weights randomly, rather than in structured patterns, aiming to reduce computational cost and memory footprint without significant accuracy loss. Current research emphasizes applying this technique to large language models and vision-language models, often employing algorithms like low-rank adapters (LoRA) and Bayesian methods to guide the pruning process and mitigate performance degradation. This area is crucial for deploying large-scale AI models on resource-constrained devices and improving the energy efficiency of AI computations.
Papers
October 15, 2024
April 16, 2024
April 3, 2024
February 27, 2024
February 25, 2024
December 12, 2023
November 28, 2023
September 19, 2023
September 12, 2023
August 9, 2023
June 9, 2023
May 3, 2023
March 18, 2023
January 2, 2023
November 25, 2022
September 15, 2022
July 1, 2022
February 17, 2022
February 9, 2022