Structural Pruning
Structural pruning aims to improve the efficiency of large neural networks, such as large language models (LLMs) and convolutional neural networks (CNNs), by removing less important parameters without significantly sacrificing performance. Current research focuses on developing novel pruning algorithms, including those incorporating reinforcement learning and optimal transport, and applying these techniques to various architectures, with a particular emphasis on LLMs and vision transformers. These advancements are significant because they enable the deployment of powerful models on resource-constrained devices and reduce the computational cost of training and inference, impacting both scientific research and practical applications.
Papers
October 15, 2024
October 11, 2024
October 9, 2024
September 17, 2024
August 27, 2024
July 24, 2024
July 19, 2024
June 22, 2024
June 18, 2024
June 15, 2024
May 3, 2024
March 28, 2024
February 12, 2024
December 8, 2023
October 26, 2023
May 19, 2023
May 18, 2023
May 3, 2023
March 17, 2023