Surrogate Gradient

Surrogate gradient methods address the challenge of training neural networks with non-differentiable components, such as the threshold functions in spiking neural networks (SNNs). Current research focuses on improving the efficiency and accuracy of surrogate gradient algorithms, exploring variations like masked surrogate gradients and forward gradient injection, and investigating their theoretical underpinnings within both deterministic and stochastic SNNs. These advancements are significant for enabling efficient training of biologically-inspired SNNs for applications like speech recognition and object detection, as well as broader applications in areas like federated learning where only zeroth-order information is available. Ultimately, refined surrogate gradient techniques aim to bridge the gap between the energy efficiency of SNNs and the performance of traditional artificial neural networks.

Papers