1. MAC과 Pooling Layer을 최적화시킨 소형 CNN 가속기 칩.
- Author
-
손현욱, 이동영, and 김형원
- Subjects
CONVOLUTIONAL neural networks ,COMPLEMENTARY metal oxide semiconductors ,INTEGERS ,MULTIPLICATION ,MEMORY - Abstract
This paper proposes a CNN accelerator which is optimized Pooling layer operation incorporated in Multiplication And Accumulation(MAC) to reduce the memory size. For optimizing memory and data path circuit, the quantized 8bit integer weights are used instead of 32bit floating-point weights for pre-training of MNIST data set. To reduce chip area, the proposed CNN model is reduced by a convolutional layer, a 4*4 Max Pooling, and two fully connected layers. And all the operations use specific MAC with approximation adders and multipliers. 94% of internal memory size reduction is achieved by simultaneously performing the convolution and the pooling operation in the proposed architecture. The proposed accelerator chip is designed by using TSMC65nmGP CMOS process. That has about half size of our previous paper, 0.8*0.9 = 0.72mm² . The presented CNN accelerator chip achieves 94% accuracy and 77us inference time per an MNIST image. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF