Back to Search Start Over

Provable distributed adaptive temporal-difference learning over time-varying networks.

Authors :
Zhu, Junlong
Li, Bing
Wang, Lin
Zhang, Mingchuan
Xing, Ling
Xi, Jiangtao
Wu, Qingtao
Source :
Expert Systems with Applications. Oct2023, Vol. 228, pN.PAG-N.PAG. 1p.
Publication Year :
2023

Abstract

Multi-agent reinforcement learning (MARL) has been successfully applied in many fields. In MARL, the policy evaluation problem is one of crucial problems. In order to solve this problem, distributed Temporal-Difference (TD) learning algorithm is one of the most popular methods in a cooperative manner. Despite its empirical success, however, the theory of the adaptive variant of distributed TD learning still remain limited. To fill this gap, we propose an adaptive distributed temporal-difference algorithm (referred to as MS - ADTD) under Markovian sampling over time-varying networks. Furthermore, we rigorously analyze the convergence of MS - ADTD , the theoretical results show that the local estimation can converge linearly to the optimal neighborhood. Meanwhile, the theoretical results are verified by simulation experiments. • This paper proposes a distributed adaptive TD learning algorithm under Markovian sampling. • It theoretically analyzes the non-asymptotic convergence performance of the proposed algorithm. • It verifies the performance of the proposed algorithm by experiments. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
09574174
Volume :
228
Database :
Academic Search Index
Journal :
Expert Systems with Applications
Publication Type :
Academic Journal
Accession number :
164285513
Full Text :
https://doi.org/10.1016/j.eswa.2023.120406