Edge Pruning
Edge pruning is a neural network compression technique aiming to reduce computational costs and memory usage by removing less important connections or parameters without significant performance degradation. Current research focuses on developing efficient pruning algorithms for various architectures, including convolutional neural networks (CNNs), vision transformers (ViTs), and large language models (LLMs), often incorporating techniques like knowledge distillation and optimization-based methods to improve performance after pruning. This work is significant because it enables the deployment of large, powerful models on resource-constrained devices and improves the energy efficiency of training and inference, impacting both scientific understanding of model redundancy and practical applications across diverse fields.
Papers
HollowNeRF: Pruning Hashgrid-Based NeRFs with Trainable Collision Mitigation
Xiufeng Xie, Riccardo Gherardi, Zhihong Pan, Stephen Huang
To prune or not to prune : A chaos-causality approach to principled pruning of dense neural networks
Rajan Sahu, Shivam Chadha, Nithin Nagaraj, Archana Mathur, Snehanshu Saha
Fantastic Weights and How to Find Them: Where to Prune in Dynamic Sparse Training
Aleksandra I. Nowak, Bram Grooten, Decebal Constantin Mocanu, Jacek Tabor
Quantifying lottery tickets under label noise: accuracy, calibration, and complexity
Viplove Arora, Daniele Irto, Sebastian Goldt, Guido Sanguinetti