Back to Search Start Over

Dynamic Programming Principles for Mean-Field Controls with Learning.

Authors :
Haotian Gu
Xin Guo
Xiaoli Wei
Renyuan Xu
Source :
Operations Research; Jul/Aug2023, Vol. 71 Issue 4, p1040-1054, 15p
Publication Year :
2023

Abstract

The dynamic programming principle (DPP) is fundamental for control and optimization, including Markov decision problems (MDPs), reinforcement learning (RL), and, more recently, mean-field controls (MFCs). However, in the learning framework of MFCs, the DPP has not been rigorously established, despite its critical importance for algorithm designs. In this paper, we first present a simple example in MFCs with learning where the DPP fails with a misspecified Q function and then propose the correct form of Q function in an appropriate space for MFCs with learning. This particular form of Q function is different from the classical one and is called the IQ function. In the special case when the transition probability and the reward are independent of the mean-field information, it integrates the classical Q function for single-agent RL over the state-action distribution. In other words, MFCs with learning can be viewed as lifting the classical RLs by replacing the state-action space with its probability distribution space. This identification of the IQ function enables us to establish precisely the DPP in the learning framework of MFCs. Finally, we illustrate through numerical experiments the time consistency of this IQ function. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
0030364X
Volume :
71
Issue :
4
Database :
Complementary Index
Journal :
Operations Research
Publication Type :
Academic Journal
Accession number :
169739722
Full Text :
https://doi.org/10.1287/opre.2022.2395