Gradient Compression
Gradient compression aims to reduce the communication overhead in distributed machine learning by transmitting smaller representations of model updates (gradients). Current research focuses on developing novel compression techniques, including quantization, sparsification, low-rank approximation, and the use of large language models as gradient priors, often incorporating error feedback mechanisms to mitigate information loss. These advancements are crucial for scaling up training of large models like LLMs and for enabling efficient federated learning in resource-constrained environments, ultimately accelerating training speed and reducing energy consumption.
Papers
November 11, 2024
October 15, 2024
October 3, 2024
September 26, 2024
July 6, 2024
July 5, 2024
July 1, 2024
May 22, 2024
May 17, 2024
February 5, 2024
February 2, 2024
January 15, 2024
December 29, 2023
December 13, 2023
December 5, 2023
November 29, 2023
November 13, 2023
June 15, 2023