Back to Search Start Over

A Sharp Estimate on the Transient Time of Distributed Stochastic Gradient Descent.

Authors :
Pu, Shi
Olshevsky, Alex
Paschalidis, Ioannis Ch.
Source :
IEEE Transactions on Automatic Control; Nov2022, Vol. 67 Issue 11, p5900-5915, 16p
Publication Year :
2022

Abstract

This article is concerned with minimizing the average of $n$ cost functions over a network, in which agents may communicate and exchange information with each other. We consider the setting where only noisy gradient information is available. To solve the problem, we study the distributed stochastic gradient descent (DSGD) method and perform a nonasymptotic convergence analysis. For strongly convex and smooth objective functions, in expectation, DSGD asymptotically achieves the optimal network-independent convergence rate compared to centralized stochastic gradient descent. Our main contribution is to characterize the transient time needed for DSGD to approach the asymptotic convergence rate. Moreover, we construct a “hard” optimization problem that proves the sharpness of the obtained result. Numerical experiments demonstrate the tightness of the theoretical results. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00189286
Volume :
67
Issue :
11
Database :
Complementary Index
Journal :
IEEE Transactions on Automatic Control
Publication Type :
Periodical
Accession number :
160621617
Full Text :
https://doi.org/10.1109/TAC.2021.3126253