Back to Search Start Over

Accelerating Deep Learning with a Parallel Mechanism Using CPU + MIC.

Authors :
Fan, Sijiang
Fei, Jiawei
Shen, Li
Source :
International Journal of Parallel Programming. Aug2018, Vol. 46 Issue 4, p660-673. 14p.
Publication Year :
2018

Abstract

Deep neural networks (DNNs) is one of the most popular machine learning methods and is widely used in many modern applications. The training process of DNNs is a time-consuming process. Accelerating the training of DNNs has been the focus of many research works. In this paper, we speed up the training of DNNs applied for automatic speech recognition and the target architecture is heterogeneous (CPU + MIC). We apply asynchronous methods for I/O and communication operations and propose an adaptive load balancing method. Besides, we also employ a momentum idea to speed up the convergence of the gradient descent algorithm. Experimental results show that our optimized algorithm is able to acquire a 20-fold speedup on a CPU + MIC platform compared with the original sequential algorithm on a single-core CPU. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
08857458
Volume :
46
Issue :
4
Database :
Academic Search Index
Journal :
International Journal of Parallel Programming
Publication Type :
Academic Journal
Accession number :
131277875
Full Text :
https://doi.org/10.1007/s10766-017-0535-9