TY - JOUR
T1 - Nice
T2 - Noise injection and clamping estimation for neural network quantization
AU - Baskin, Chaim
AU - Zheltonozhkii, Evgenii
AU - Rozen, Tal
AU - Liss, Natan
AU - Chai, Yoav
AU - Schwartz, Eli
AU - Giryes, Raja
AU - Bronstein, Alexander M.
AU - Mendelson, Avi
N1 - Publisher Copyright: © 2021 by the authors. Licensee MDPI, Basel, Switzerland.
PY - 2021/9/1
Y1 - 2021/9/1
N2 - Convolutional Neural Networks (CNNs) are very popular in many fields including computer vision, speech recognition, natural language processing, etc. Though deep learning leads to groundbreaking performance in those domains, the networks used are very computationally de-manding and are far from being able to perform in real-time applications even on a GPU, which is not power efficient and therefore does not suit low power systems such as mobile devices. To overcome this challenge, some solutions have been proposed for quantizing the weights and activations of these networks, which accelerate the runtime significantly. Yet, this acceleration comes at the cost of a larger error unless spatial adjustments are carried out. The method proposed in this work trains quantized neural networks by noise injection and a learned clamping, which improve accuracy. This leads to state-of-the-art results on various regression and classification tasks, e.g., ImageNet classification with architectures such as ResNet-18/34/50 with as low as 3 bit weights and activations. We implement the proposed solution on an FPGA to demonstrate its applicability for low-power real-time applications. The quantization code will become publicly available upon acceptance.
AB - Convolutional Neural Networks (CNNs) are very popular in many fields including computer vision, speech recognition, natural language processing, etc. Though deep learning leads to groundbreaking performance in those domains, the networks used are very computationally de-manding and are far from being able to perform in real-time applications even on a GPU, which is not power efficient and therefore does not suit low power systems such as mobile devices. To overcome this challenge, some solutions have been proposed for quantizing the weights and activations of these networks, which accelerate the runtime significantly. Yet, this acceleration comes at the cost of a larger error unless spatial adjustments are carried out. The method proposed in this work trains quantized neural networks by noise injection and a learned clamping, which improve accuracy. This leads to state-of-the-art results on various regression and classification tasks, e.g., ImageNet classification with architectures such as ResNet-18/34/50 with as low as 3 bit weights and activations. We implement the proposed solution on an FPGA to demonstrate its applicability for low-power real-time applications. The quantization code will become publicly available upon acceptance.
KW - CNN architecture
KW - Low power
KW - Neural networks
KW - Quantization
UR - http://www.scopus.com/inward/record.url?scp=85114454490&partnerID=8YFLogxK
U2 - https://doi.org/10.3390/math9172144
DO - https://doi.org/10.3390/math9172144
M3 - Article
SN - 2227-7390
VL - 9
JO - Mathematics
JF - Mathematics
IS - 17
M1 - 2144
ER -