Back to Search Start Over

Exponential Discretization of Weights of Neural Network Connections in Pre-Trained Neural Networks.

Authors :
Malsagov, M. Yu.
Khayrov, E. M.
Pushkareva, M. M.
Karandashev, I. M.
Source :
Optical Memory & Neural Networks; Oct2019, Vol. 28 Issue 4, p262-270, 9p
Publication Year :
2019

Abstract

To reduce random access memory (RAM) requirements and to increase speed of recognition algorithms we consider a weight discretization problem for trained neural networks. We show that an exponential discretization is preferable to a linear discretization since it allows one to achieve the same accuracy when the number of bits is 1 or 2 less. The quality of the neural network VGG-16 is already satisfactory (top5 accuracy 69%) in the case of 3 bit exponential discretization. The ResNet50 neural network shows top5 accuracy 84% at 4 bits. Other neural networks perform fairly well at 5 bits (top5 accuracies of Xception, Inception-v3, and MobileNet-v2 top5 were 87%, 90%, and 77%, respectively). At less number of bits, the accuracy decreases rapidly. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
1060992X
Volume :
28
Issue :
4
Database :
Complementary Index
Journal :
Optical Memory & Neural Networks
Publication Type :
Academic Journal
Accession number :
141662919
Full Text :
https://doi.org/10.3103/S1060992X19040106