Back to Search Start Over

A Modified Nonlinear Conjugate Gradient Algorithm for Large-Scale Nonsmooth Convex Optimization.

Authors :
Woldu, Tsegay Giday
Zhang, Haibin
Zhang, Xin
Fissuh, Yemane Hailu
Source :
Journal of Optimization Theory & Applications; Apr2020, Vol. 185 Issue 1, p223-238, 16p
Publication Year :
2020

Abstract

Nonlinear conjugate gradient methods are among the most preferable and effortless methods to solve smooth optimization problems. Due to their clarity and low memory requirements, they are more desirable for solving large-scale smooth problems. Conjugate gradient methods make use of gradient and the previous direction information to determine the next search direction, and they require no numerical linear algebra. However, the utility of nonlinear conjugate gradient methods has not been widely employed in solving nonsmooth optimization problems. In this paper, a modified nonlinear conjugate gradient method, which achieves the global convergence property and numerical efficiency, is proposed to solve large-scale nonsmooth convex problems. The new method owns the search direction, which generates sufficient descent property and belongs to a trust region. Under some suitable conditions, the global convergence of the proposed algorithm is analyzed for nonsmooth convex problems. The numerical efficiency of the proposed algorithm is tested and compared with some existing methods on some large-scale nonsmooth academic test problems. The numerical results show that the new algorithm has a very good performance in solving large-scale nonsmooth problems. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00223239
Volume :
185
Issue :
1
Database :
Complementary Index
Journal :
Journal of Optimization Theory & Applications
Publication Type :
Academic Journal
Accession number :
142491101
Full Text :
https://doi.org/10.1007/s10957-020-01636-7