Back to Search Start Over

Hierarchical state-abstracted and socially augmented Q-Learning for reducing complexity in agent-based learning.

Authors :
Sun, Xueqing
Mao, Tao
Ray, Laura
Shi, Dongqing
Kralik, Jerald
Source :
Journal of Control Theory & Applications; Aug2011, Vol. 9 Issue 3, p440-450, 11p
Publication Year :
2011

Abstract

primary challenge of agent-based policy learning in complex and uncertain environments is escalating computational complexity with the size of the task space (action choices and world states) and the number of agents. Nonetheless, there is ample evidence in the natural world that high-functioning social mammals learn to solve complex problems with ease, both individually and cooperatively. This ability to solve computationally intractable problems stems from both brain circuits for hierarchical representation of state and action spaces and learned policies as well as constraints imposed by social cognition. Using biologically derived mechanisms for state representation and mammalian social intelligence, we constrain state-action choices in reinforcement learning in order to improve learning efficiency. Analysis results bound the reduction in computational complexity due to state abstraction, hierarchical representation, and socially constrained action selection in agent-based learning problems that can be described as variants of Markov decision processes. Investigation of two task domains, single-robot herding and multirobot foraging, shows that theoretical bounds hold and that acceptable policies emerge, which reduce task completion time, computational cost, and/or memory resources compared to learning without hierarchical representations and with no social knowledge. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
16726340
Volume :
9
Issue :
3
Database :
Complementary Index
Journal :
Journal of Control Theory & Applications
Publication Type :
Academic Journal
Accession number :
62867587
Full Text :
https://doi.org/10.1007/s11768-011-1047-6