Back to Search Start Over

Enhancing Policy Gradient with the Polyak Step-Size Adaption

Authors :
Li, Yunxiang
Yuan, Rui
Fan, Chen
Schmidt, Mark
Horváth, Samuel
Gower, Robert M.
Takáč, Martin
Publication Year :
2024

Abstract

Policy gradient is a widely utilized and foundational algorithm in the field of reinforcement learning (RL). Renowned for its convergence guarantees and stability compared to other RL algorithms, its practical application is often hindered by sensitivity to hyper-parameters, particularly the step-size. In this paper, we introduce the integration of the Polyak step-size in RL, which automatically adjusts the step-size without prior knowledge. To adapt this method to RL settings, we address several issues, including unknown f* in the Polyak step-size. Additionally, we showcase the performance of the Polyak step-size in RL through experiments, demonstrating faster convergence and the attainment of more stable policies.

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2404.07525
Document Type :
Working Paper