Back to Search Start Over

Integrating human learning and reinforcement learning: A novel approach to agent training.

Authors :
Li, Yao-Hui
Zhang, Feng
Hua, Qiang
Zhou, Xiao-Hua
Source :
Knowledge-Based Systems. Jun2024, Vol. 294, pN.PAG-N.PAG. 1p.
Publication Year :
2024

Abstract

Off-policy reinforcement learning (RL) algorithms are known for improving sample efficiency by employing prior experiences in experience replay memory. However, most existing off-policy RL algorithms are bottlenecked by the slow convergence speed and demand of a large number of interaction samples. In contrast, on-policy RL algorithms converge fast and continuously generate new samples. However, their success heavily depends on the accuracy of the generated samples. To address these challenges, a novel RL framework called HI-FER (H uman-learning I nspired F requent E xperience R eplay) is proposed by mimicking the process of human learning. HI-FER employs a parallelized experience replay and repetitive training framework to expedite the convergence rate of off-policy algorithms, which imitates the function of the human brain's parallel information processing and repetitive learning. Additionally, a periodic network reset strategy and dynamic memory updating are leveraged by imitating the forgetting mechanism of humans to prevent overfitting triggered by repetitive updating on limited experiences. Extensive comparison experiments and ablation studies are performed on benchmark environments to evaluate the proposed method. The empirical results demonstrate that HI-FER outperforms the baselines in terms of sample efficiency on state-based (14% improvements) and image-based (51% improvements) tasks from DMControl. Project website and code: https://github.com/Arya87/HI-FER. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
09507051
Volume :
294
Database :
Academic Search Index
Journal :
Knowledge-Based Systems
Publication Type :
Academic Journal
Accession number :
177088958
Full Text :
https://doi.org/10.1016/j.knosys.2024.111782