Residual Momentum
Residual momentum, a technique enhancing optimization algorithms like stochastic gradient descent (SGD), aims to improve convergence speed and generalization performance in various machine learning tasks. Current research focuses on adapting momentum methods for federated learning, addressing challenges in asynchronous settings and non-convex optimization problems, and exploring its impact on model architectures ranging from linear networks to deep neural networks and transformers. These advancements are significant because they lead to more efficient training of large-scale models and improved performance in applications such as image segmentation, natural language processing, and even robotic control.
Papers
March 8, 2024
February 19, 2024
February 15, 2024
February 4, 2024
January 22, 2024
December 23, 2023
November 25, 2023
November 6, 2023
October 24, 2023
October 9, 2023
September 13, 2023
July 27, 2023
July 20, 2023
June 28, 2023
June 15, 2023
June 13, 2023
May 28, 2023
April 14, 2023
March 31, 2023