Back to Search Start Over

Skill enhancement learning with knowledge distillation.

Authors :
Liu, Naijun
Sun, Fuchun
Fang, Bin
Liu, Huaping
Source :
SCIENCE CHINA Information Sciences; Aug2024, Vol. 67 Issue 8, p1-15, 15p
Publication Year :
2024

Abstract

Skill learning through reinforcement learning has significantly progressed in recent years. However, it often struggles to efficiently find optimal or near-optimal policies due to the inherent trial-and-error exploration in reinforcement learning. Although algorithms have been proposed to enhance skill learning efficacy, there is still much room for improvement in terms of skill learning performance and training stability. In this paper, we propose an algorithm called skill enhancement learning with knowledge distillation (SELKD), which integrates multiple actors and multiple critics for skill learning. SELKD employs knowledge distillation to establish a mutual learning mechanism among actors. To mitigate critic overestimation bias, we introduce a novel target value calculation method. We also perform theoretical analysis to ensure the convergence of SELKD. Finally, experiments are conducted on several continuous control tasks, illustrating the effectiveness of the proposed algorithm. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
1674733X
Volume :
67
Issue :
8
Database :
Complementary Index
Journal :
SCIENCE CHINA Information Sciences
Publication Type :
Academic Journal
Accession number :
178671141
Full Text :
https://doi.org/10.1007/s11432-023-4016-0