Quantized Neural Network
Quantized neural networks (QNNs) aim to reduce the computational cost and memory footprint of deep learning models by representing weights and activations using lower-precision integer arithmetic, rather than 32-bit floating-point numbers. Current research focuses on improving the accuracy of QNNs through techniques like quantization-aware training, exploring different quantization schemes (e.g., mixed-precision, stochastic quantization), and developing efficient algorithms for training and verification. This field is significant because QNNs enable the deployment of deep learning on resource-constrained devices, impacting applications ranging from mobile and edge computing to embedded systems and Internet of Things (IoT) devices.
Papers
October 7, 2024
August 30, 2024
April 9, 2024
March 22, 2024
December 20, 2023
December 14, 2023
November 30, 2023
November 6, 2023
October 29, 2023
September 25, 2023
July 29, 2023
July 8, 2023
June 23, 2023
June 16, 2023
May 25, 2023
March 15, 2023
December 29, 2022
December 10, 2022
November 29, 2022