Parameter Pruning
Parameter pruning aims to reduce the size and computational cost of large neural networks, like large language models (LLMs) and convolutional neural networks (CNNs), without significantly sacrificing performance. Current research focuses on developing efficient pruning algorithms, including those based on magnitude, layer merging, and concurrent training-pruning, often applied to specific architectures like Transformers and ResNets. These advancements are crucial for deploying complex models on resource-constrained devices and mitigating the environmental impact of large-scale training, impacting both scientific research and practical applications in various fields.
Papers
October 12, 2024
June 24, 2024
June 17, 2024
June 6, 2024
January 15, 2024
October 6, 2023
June 22, 2023
March 2, 2023
February 17, 2023
September 29, 2022
July 8, 2022
May 25, 2022