Back to Search
Start Over
A single timescale stochastic quasi-Newton method for stochastic optimization.
- Source :
- International Journal of Computer Mathematics; Dec2023, Vol. 100 Issue 12, p2196-2216, 21p
- Publication Year :
- 2023
-
Abstract
- In this paper, we propose a single timescale stochastic quasi-Newton method for solving the stochastic optimization problems. The objective function of the problem is a composition of two smooth functions and their derivatives are not available. The algorithm sets to approximate sequences to estimate the gradient of the composite objective function and the inner function. The matrix correction parameters are given in BFGS update form for avoiding the assumption that Hessian matrix of objective is positive definite. We show the global convergence of the algorithm. The algorithm achieves the complexity O (ϵ − 1) to find an ϵ − approximate stationary point and ensure that the expectation of the squared norm of the gradient is smaller than the given accuracy tolerance ϵ. The numerical results of nonconvex binary classification problem using the support vector machine and a multicall classification problem using neural networks are reported to show the effectiveness of the algorithm. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 00207160
- Volume :
- 100
- Issue :
- 12
- Database :
- Complementary Index
- Journal :
- International Journal of Computer Mathematics
- Publication Type :
- Academic Journal
- Accession number :
- 173686800
- Full Text :
- https://doi.org/10.1080/00207160.2023.2269430