Back to Search Start Over

Laius: an energy-efficient FPGA CNN accelerator with the support of a fixed-point training framework

Authors :
Nie, Zikai
Li, Zhisheng
Wang, Lei
Guo, Shasha
Deng, Yu
Deng, Rangyu
Dou, Qiang
Source :
International Journal of Computational Science and Engineering; 2020, Vol. 21 Issue: 3 p418-428, 11p
Publication Year :
2020

Abstract

With the development of convolutional neural networks (CNNs), their high computational complexity and energy consumption become significant problems. Many CNN inference accelerators are proposed to reduce the consumption. Most of them are based on 32-bit float-point matrix multiplication, where the data precision is over-provisioned. This paper presents Laius, an 8-bit fixed-point LeNet inference engine implemented on FPGA. To achieve low-precision computation and storage, we introduce our fixed-point training framework called FixCaffe. To economise FPGA resources, we proposed a methodology to find the optimal bit-length for weight and bias in LeNet. We use optimisations of pipelining, tiling, and theoretical analysis to improve the performance. Experiment results show that Laius achieves 44.9 Gops throughputs. Moreover, with only 1% accuracy loss, 8-bit Laius largely reduces 31.43% in delay, 87.01% in LUT consumption, 66.50% in BRAM consumption, 65.11% in DSP consumption and 47.95% in power compared to the 32-bit version with the same structure.

Details

Language :
English
ISSN :
17427185 and 17427193
Volume :
21
Issue :
3
Database :
Supplemental Index
Journal :
International Journal of Computational Science and Engineering
Publication Type :
Periodical
Accession number :
ejs52799327
Full Text :
https://doi.org/10.1504/IJCSE.2020.106064