Back to Search Start Over

Multi-core parallel architecture design and experiment for deep learning model training.

Authors :
Wanwu, Li
Lin, Liu
Jixian, Zhang
Shuai, Liu
Jiahao, Qiu
Source :
Multimedia Tools & Applications; Mar2022, Vol. 81 Issue 8, p11587-11604, 18p
Publication Year :
2022

Abstract

The parallel architecture can improve the training speed of Deep learning models and it is beneficial to model optimization. This research designs Optimal Interleaved Distributed Architecture (OIDA). Its characteristics are (1) the data set that each child node participates in training is fixed and unique; (2) multiple approaches ensure that the model parameters of each child node participating in training each time are always optimal; (3) the computing units of each child node are interleaved for training without stopping. The architecture can be deployed as Single Machine Parallel (SMP) or Multi Machine Distributed (MMD), or a combination of both. The experimental results show that its efficiency is obvious in a multi-machine multi-card environment, and the model speed of training millions of parameters in a 5-machine 10-card lightweight (Graphics Processing Unit (GPU) of 2GB memory) cluster can reach megapixels/s. When training the same amount of training data, GPU can process data faster than Central Processing Unit (CPU) with the same number of clusters. 4-machine 4-core GPU parallel saves 65.61% of time compared with 4-machine 4-core CPU parallel, while 4-machine 8-core GPU saves 83.32% of time compared with 4-machine 4-core CPU parallel. For polarized Synthetic Aperture Radar (SAR) Deep learning data sets, increasing the number of computers in the cluster can effectively save training time. In the experimental environment of this research, there is no situation that when the number of computers in the cluster increases due to network data transmission, the training time does not decrease. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
13807501
Volume :
81
Issue :
8
Database :
Complementary Index
Journal :
Multimedia Tools & Applications
Publication Type :
Academic Journal
Accession number :
156025387
Full Text :
https://doi.org/10.1007/s11042-022-12292-6