Stochastic Quantization
Stochastic quantization (SQ) is a technique used to reduce the computational and communication costs associated with large-scale machine learning models by representing model parameters with fewer bits. Current research focuses on applying SQ to improve the efficiency and robustness of various algorithms, including federated learning, clustering, and deep neural network training, often incorporating adaptive quantization schemes and differentiable quantizers to mitigate accuracy loss. This work is significant because it addresses critical bottlenecks in deploying large models, enabling faster training, reduced communication overhead, and enhanced privacy in distributed settings.
Papers
September 3, 2024
May 22, 2024
March 15, 2024
December 10, 2023
November 30, 2023
November 6, 2023
September 29, 2023
September 20, 2023
March 11, 2023
December 12, 2022
June 9, 2022
May 16, 2022
May 9, 2022
March 11, 2022