Binarization Method
Binarization methods aim to drastically reduce the computational cost and memory footprint of neural networks by representing weights and/or activations using only one bit. Current research focuses on improving the accuracy of binarized models, particularly for large language models (LLMs) and vision transformers (ViTs), through techniques like alternating refined binarization, learnable binarization during training, and incorporating structural sparsity. These advancements are significant because they enable the deployment of powerful deep learning models on resource-constrained devices, impacting diverse fields from natural language processing and computer vision to mobile and embedded systems.
Papers
November 15, 2024
October 4, 2024
August 6, 2024
August 3, 2024
June 18, 2024
June 9, 2024
May 23, 2024
May 22, 2024
April 8, 2024
April 1, 2024
March 29, 2024
February 27, 2024
February 20, 2024
February 19, 2024
January 22, 2024
December 4, 2023
December 1, 2023
October 2, 2023