Bit Training
Bit training focuses on reducing the numerical precision used to represent neural network weights and activations, aiming to accelerate training and reduce memory consumption without significant accuracy loss. Current research explores various bit-widths (4-bit, 8-bit, 16-bit), often employing techniques like quantization and specialized arithmetic to maintain performance, with applications to transformer models and large language models (LLMs) being particularly prominent. This approach holds significant promise for making deep learning more efficient and accessible, enabling training of larger models on resource-constrained hardware and facilitating faster model updates.
Papers
November 30, 2023
August 12, 2023
June 21, 2023
May 23, 2023
November 19, 2022