Stale Gradient
Stale gradient problems arise in distributed machine learning when model updates are based on outdated information, hindering efficient and accurate training. Current research focuses on mitigating this issue through techniques like corrector networks that adjust stale embeddings, asynchronous mini-batching strategies that handle variable delays, and adaptive algorithms that selectively utilize fresh and stale updates in federated learning and other distributed settings. Addressing stale gradients is crucial for improving the scalability and efficiency of large-scale machine learning, impacting diverse applications from recommendation systems to federated learning across heterogeneous devices.
Papers
September 3, 2024
August 14, 2024
May 7, 2024
March 22, 2024
December 18, 2023
December 16, 2023
August 29, 2023
August 25, 2023
April 14, 2023
February 2, 2023
November 2, 2022
June 18, 2022
May 31, 2022