Network Pruning
Network pruning aims to reduce the size and computational cost of deep neural networks (DNNs) without significant performance loss, primarily by removing less important weights or connections. Current research focuses on developing efficient pruning algorithms for large language models (LLMs), convolutional neural networks (CNNs), and spiking neural networks (SNNs), often employing techniques like structured or unstructured pruning, and incorporating optimization methods to improve accuracy and speed. These advancements are crucial for deploying large-scale DNNs on resource-constrained devices, improving energy efficiency, and accelerating inference times across various applications.
Papers
November 14, 2024
November 11, 2024
October 19, 2024
October 18, 2024
October 9, 2024
September 30, 2024
September 27, 2024
September 20, 2024
August 26, 2024
August 22, 2024
June 7, 2024
June 3, 2024
May 23, 2024
April 9, 2024
March 21, 2024
March 19, 2024
March 11, 2024
January 16, 2024