Back to Search Start Over

Powered stochastic optimization with hypergradient descent for large-scale learning systems.

Authors :
Yang, Zhuang
Li, Xiaotian
Source :
Expert Systems with Applications. Mar2024:Part C, Vol. 238, pN.PAG-N.PAG. 1p.
Publication Year :
2024

Abstract

Stochastic optimization (SO) algorithms based on the Powerball function, namely powered stochastic optimization (PoweredSO) algorithms, have been confirmed, effectively, and demonstrated great potential in the context of large-scale optimization and machine learning tasks. Nevertheless, the issue of how to determine the learning rate for PoweredSO is a challenge and still unsolved problem. In this paper, we propose a class of adaptive PoweredSO approaches that are efficient, scalable and robust. It takes advantage of the hypergradient descent (HD) technique to automatically acquire an online learning rate for PoweredSO-like methods. In the first part, we study the behavior of the canonical PoweredSO algorithm, the Powerball stochastic gradient descent (pbSGD) method, with HD. The existing PoweredSO algorithms also suffer from the high variance because they take the similar algorithmic framework to SO algorithms, arising from sampling tactics. Therefore, the second portion develops an adaptive powered variance-reduced optimization method via utilizing both variance-reduced technique and HD. Moreover, we present the convergence analysis of the proposed algorithms and explore their iteration complexity on non-convex cases. Numerical experiments are conducted on machine learning tasks, verifying the superior performance over modern SO algorithms. • We propose a type of adaptive powered stochastic optimization algorithms. • The theoretical analysis of the resulting algorithms is established. • Experiments on machine learning tasks verify the efficacy of the algorithms. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
09574174
Volume :
238
Database :
Academic Search Index
Journal :
Expert Systems with Applications
Publication Type :
Academic Journal
Accession number :
173706017
Full Text :
https://doi.org/10.1016/j.eswa.2023.122017