Forward Gradient
Forward gradient methods aim to compute gradients for training neural networks without the backward pass of backpropagation, addressing its memory limitations and computational cost. Current research focuses on improving the accuracy and efficiency of forward gradient estimation, exploring techniques like variance reduction, local loss functions, and integration with specific architectures such as invertible networks and MLPMixer-inspired designs. These advancements offer potential for faster training and reduced memory consumption in various applications, including computational fluid dynamics, federated learning of large language models, and quantum chemistry simulations.
Papers
November 6, 2024
October 23, 2024
May 31, 2024
March 19, 2024
February 22, 2024
February 15, 2024
November 20, 2023
August 26, 2023
July 17, 2023
June 12, 2023
December 19, 2022
December 14, 2022
October 7, 2022
September 13, 2022