Back to Search Start Over

UNIQ: Uniform Noise Injection for Non-Uniform Quantization of Neural Networks.

Authors :
BASKIN, CHAIM
LISS, NATAN
SCHWARTZ, ELI
ZHELTONOZHSKII, EVGENII
GIRYES, RAJA
BRONSTEIN, ALEX M.
MENDELSON, AVI
Source :
ACM Transactions on Computer Systems; Jan2021, Vol. 37 Issue 1-4, p1-15, 15p
Publication Year :
2021

Abstract

We present a novel method for neural network quantization. Our method, named UNIQ, emulates a nonuniform k-quantile quantizer and adapts themodel to perform well with quantized weights by injecting noise to the weights at training time. As a by-product of injecting noise to weights, we find that activations can also be quantized to as low as 8-bit with only a minor accuracy degradation. Our non-uniform quantization approach provides a novel alternative to the existing uniform quantization techniques for neural networks. We further propose a novel complexity metric of number of bit operations performed (BOPs), and we show that this metric has a linear relation with logic utilization and power. We suggest evaluating the trade-off of accuracy vs. complexity (BOPs). The proposed method, when evaluated on ResNet18/34/50 and MobileNet on ImageNet, outperforms the prior state of the art both in the low-complexity regime and the high accuracy regime. We demonstrate the practical applicability of this approach, by implementing our non-uniformly quantized CNN on FPGA. [ABSTRACT FROM AUTHOR]

Subjects

Subjects :
WEIGHT training
NOISE
DEEP learning

Details

Language :
English
ISSN :
07342071
Volume :
37
Issue :
1-4
Database :
Complementary Index
Journal :
ACM Transactions on Computer Systems
Publication Type :
Academic Journal
Accession number :
151373991
Full Text :
https://doi.org/10.1145/3444943