Quantization Robust Parameter
Quantization robust parameters aim to create neural network models that maintain accuracy even when their parameters are reduced to lower precision (e.g., 8-bit or even 2-bit integers), crucial for deploying models on resource-constrained devices. Current research focuses on developing methods to predict these robust parameters, often employing graph hypernetworks or incorporating quantization-aware training strategies to improve model resilience. This work is significant because it addresses the trade-off between model size/speed and accuracy, enabling efficient deployment of deep learning in various applications while mitigating the vulnerabilities introduced by quantization.
Papers
September 24, 2023
August 4, 2023
April 8, 2023
October 17, 2022
August 26, 2022
July 31, 2022
March 11, 2022