Parameter Efficient Sparsity
Parameter-efficient sparsity focuses on reducing the computational cost and memory footprint of large neural networks, particularly in vision transformers (ViTs) and large language models (LLMs), without significant performance degradation. Current research emphasizes post-training sparsity methods, exploring techniques like blockwise pruning, differentiable sparsity allocation, and optimized matrix multiplication kernels to achieve this. These advancements are crucial for deploying large models on resource-constrained devices and improving training efficiency, impacting both the scalability of AI research and the accessibility of powerful models for various applications.
Papers
November 4, 2024
October 9, 2024
May 9, 2024
February 21, 2024
February 18, 2024
January 16, 2024
January 5, 2024
October 23, 2023
September 15, 2023
August 3, 2023
July 8, 2023
July 7, 2023
March 10, 2023
January 19, 2023
January 17, 2023
November 28, 2022
May 23, 2022
April 2, 2022
March 9, 2022