Sparse Backpropagation
Sparse backpropagation aims to accelerate and reduce the memory footprint of training deep neural networks by selectively updating only a subset of model parameters during backpropagation. Current research focuses on developing dynamic sparsity algorithms, such as TinyProp and its variants, and applying sparse backpropagation to various architectures, including Mixture-of-Experts (MoE) models and recurrent neural networks. These advancements are particularly impactful for resource-constrained environments like edge devices and embedded systems, enabling on-device learning and fine-tuning of large models while maintaining accuracy. The resulting efficiency gains are significant for both training speed and energy consumption.
Papers
September 11, 2024
October 26, 2023
October 1, 2023
August 17, 2023
February 9, 2023
June 13, 2022