Back to Search Start Over

An In-Memory-Computing Binary Neural Network Architecture With In-Memory Batch Normalization

Authors :
Prathamesh Prashant Rege
Ming Yin
Sanjay Parihar
Joseph Versaggi
Shashank Nemawarkar
Source :
IEEE Access, Vol 12, Pp 190889-190896 (2024)
Publication Year :
2024
Publisher :
IEEE, 2024.

Abstract

This paper describes an in-memory computing architecture that combines full-precision computation for the first and last layers of a neural network while employing binary weights and input activations for the intermediate layers. This unique approach presents an efficient and effective solution for optimizing neural-network computations, reducing complexity, and enhancing energy efficiency. Notably, multiple architecture-level optimization methods are developed to ensure the binary operations thereby eliminating the need for intricate “digital logic” components external to the memory units. One of the key contributions of this study is in-memory batch normalization, which is implemented to provide good accuracy for CIFAR10 classification applications. Despite the inherent challenges posed by the process variations, the proposed design demonstrated an accuracy of 78%. Furthermore, the SRAM layer in the architecture showed an energy efficiency of 1086 TOPS/W and throughput of 23 TOPS, all packed efficiently within an area of 60 TOPS/mm2. This novel in-memory computing architecture offers a promising solution for next-generation efficient and high-performance deep learning applications.

Details

Language :
English
ISSN :
21693536
Volume :
12
Database :
Directory of Open Access Journals
Journal :
IEEE Access
Publication Type :
Academic Journal
Accession number :
edsdoj.771dfe3f27e6483293c536acc86afa1d
Document Type :
article
Full Text :
https://doi.org/10.1109/ACCESS.2024.3444481