Residual Momentum
Residual momentum, a technique enhancing optimization algorithms like stochastic gradient descent (SGD), aims to improve convergence speed and generalization performance in various machine learning tasks. Current research focuses on adapting momentum methods for federated learning, addressing challenges in asynchronous settings and non-convex optimization problems, and exploring its impact on model architectures ranging from linear networks to deep neural networks and transformers. These advancements are significant because they lead to more efficient training of large-scale models and improved performance in applications such as image segmentation, natural language processing, and even robotic control.
Papers
March 1, 2023
December 29, 2022
December 14, 2022
November 17, 2022
November 10, 2022
November 2, 2022
October 28, 2022
October 12, 2022
September 30, 2022
September 27, 2022
September 8, 2022
August 11, 2022
August 3, 2022
July 13, 2022
June 29, 2022
June 8, 2022
March 23, 2022
March 22, 2022
January 28, 2022