Back to Search Start Over

Efficient MPI‐AllReduce for large‐scale deep learning on GPU‐clusters.

Authors :
Thao Nguyen, Truong
Wahib, Mohamed
Takano, Ryousei
Source :
Concurrency & Computation: Practice & Experience; 6/25/2021, Vol. 33 Issue 12, p1-20, 20p
Publication Year :
2021

Abstract

Summary: Training models on large‐scale GPUs‐accelerated clusters are becoming a commonplace due to the increase in complexity and size in deep learning models. One of the main challenges for distributed training is the collective communication overhead for large message sizes: up to hundreds of MB. In this paper, we propose two hierarchical distributed memory multileader AllReduce algorithms optimized for GPU‐accelerated clusters (named lr_lr and lr_rab), in which GPUs inside a computing node perform an intra‐node communication phase to gather and store results of local reduced values to designated GPUs (known as node leaders). Node leaders then keep a role as an inter‐node communicator. Each leader exchanges one part of reduced values to the leaders of the other nodes in parallel. Hence, we are capable of significantly reducing the time for injecting data into the inter‐node network. We also overlap the inter‐node and intra‐node communication by implementing our proposal in a pipelined manner. We evaluate those algorithms on the discrete‐event simulation Simgrid. We show that our algorithms, lr_lr and lr_rab, can cut down the execution time of an AllReduce microbenchmark that uses the logical ring algorithm (lr) by up to 45% and 51%, respectively. With the pipelined implementation, our lr_lr_pipe achieves 15% performance improvement when compared with lr_lr. In addition, the simulation result also projects power savings for the network devices of up to 23% and 32%. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
15320626
Volume :
33
Issue :
12
Database :
Complementary Index
Journal :
Concurrency & Computation: Practice & Experience
Publication Type :
Academic Journal
Accession number :
150473637
Full Text :
https://doi.org/10.1002/cpe.5574