Back to Search Start Over

CBin-NN: An Inference Engine for Binarized Neural Networks.

Authors :
Sakr, Fouad
Berta, Riccardo
Doyle, Joseph
Capello, Alessio
Dabbous, Ali
Lazzaroni, Luca
Bellotti, Francesco
Source :
Electronics (2079-9292); May2024, Vol. 13 Issue 9, p1624, 17p
Publication Year :
2024

Abstract

Binarization is an extreme quantization technique that is attracting research in the Internet of Things (IoT) field, as it radically reduces the memory footprint of deep neural networks without a correspondingly significant accuracy drop. To support the effective deployment of Binarized Neural Networks (BNNs), we propose CBin-NN, a library of layer operators that allows the building of simple yet flexible convolutional neural networks (CNNs) with binary weights and activations. CBin-NN is platform-independent and is thus portable to virtually any software-programmable device. Experimental analysis on the CIFAR-10 dataset shows that our library, compared to a set of state-of-the-art inference engines, speeds up inference by 3.6 times and reduces the memory required to store model weights and activations by 7.5 times and 28 times, respectively, at the cost of slightly lower accuracy (2.5%). An ablation study stresses the importance of a Quantized Input Quantized Kernel Convolution layer to improve accuracy and reduce latency at the cost of a slight increase in model size. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
20799292
Volume :
13
Issue :
9
Database :
Complementary Index
Journal :
Electronics (2079-9292)
Publication Type :
Academic Journal
Accession number :
177180062
Full Text :
https://doi.org/10.3390/electronics13091624