Pruned Model
Pruned models aim to reduce the size and computational cost of deep neural networks (DNNs) by removing less important parameters or neurons, while preserving accuracy. Current research focuses on developing efficient pruning algorithms, including structured pruning methods for large language models (LLMs) and vision transformers, and exploring techniques like optimal transport for mitigating security vulnerabilities in pruned models. This work is significant because it addresses the growing need for deploying DNNs on resource-constrained devices and improves the efficiency and security of various machine learning applications.
Papers
October 18, 2024
August 28, 2024
July 23, 2024
July 18, 2024
June 7, 2024
February 26, 2024
February 8, 2024
September 20, 2023
September 13, 2023
August 17, 2023
June 16, 2023
May 28, 2023
January 27, 2023
January 26, 2023