Back to Search Start Over

Data and Hardware Efficient Design for Convolutional Neural Network.

Authors :
Lin, Yue-Jin
Chang, Tian Sheuan
Source :
IEEE Transactions on Circuits & Systems. Part I: Regular Papers; May2018, Vol. 65 Issue 5, p1642-1651, 10p
Publication Year :
2018

Abstract

Hardware design of deep convolutional neural networks (CNNs) faces challenges of high computational complexity and data bandwidth as well as huge divergence in different CNN network layers, in which the throughput of the convolutional layer would be bounded by available hardware resource, and throughput of the fully connected layer would be bounded by available data bandwidth. Thus, a highly flexible and efficient design is desired to meet these needs. This paper presents an end-to-end CNN accelerator that maximizes hardware utilization with run-time configurations of different kernel sizes. It also minimizes data bandwidth with the output first strategy to improve the data reuse of the convolutional layers by up to $300\times \sim 600\times $ compared with the non-reused case. The whole CNN implementation of the target network is generated optimally for both hardware and data efficiency under design resource constraints, which can be run-time reconfigured by the layer optimized parameters to achieve real-time and end-to-end CNN acceleration. An implementation example for AlexNet consumes a 1.783 M gate count for 216 MACs and a 142.64 kb internal buffer with TSMC 40 nm process, and achieves 99.7 and 61.6 f/s under 454 MHz clock frequency for the convolutional layers and the whole AlexNet, respectively. [ABSTRACT FROM PUBLISHER]

Details

Language :
English
ISSN :
15498328
Volume :
65
Issue :
5
Database :
Complementary Index
Journal :
IEEE Transactions on Circuits & Systems. Part I: Regular Papers
Publication Type :
Periodical
Accession number :
128843404
Full Text :
https://doi.org/10.1109/TCSI.2017.2759803