Back to Search Start Over

A SMOOTHING PROXIMAL GRADIENT ALGORITHM FOR NONSMOOTH CONVEX REGRESSION WITH CARDINALITY PENALTY.

Authors :
WEI BIAN
XIAOJUN CHEN
Source :
SIAM Journal on Numerical Analysis; 2020, Vol. 58 Issue 1, p858-883, 26p
Publication Year :
2020

Abstract

In this paper, we focus on the constrained sparse regression problem, where the loss function is convex but nonsmooth and the penalty term is defined by the cardinality function. First, we give an exact continuous relaxation problem in the sense that both problems have the same optimal solution set. Moreover, we show that a vector is a local minimizer with the lower bound property of the original problem if and only if it is a lifted stationary point of the relaxation problem. Second, we propose a smoothing proximal gradient (SPG) algorithm for finding a lifted stationary point of the continuous relaxation model. Our algorithm is a novel combination of the classical proximal gradient algorithm and the smoothing method. We prove that the proposed SPG algorithm globally converges to a lifted stationary point of the relaxation problem, has the local convergence rate of o(k \tau) with \tau \in (0, 1 2) on the objective function value, and identifies the zero entries of the lifted stationary point in finite iterations. Finally, we use three examples to illustrate the validity of the continuous relaxation model and good numerical performance of the SPG algorithm. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00361429
Volume :
58
Issue :
1
Database :
Complementary Index
Journal :
SIAM Journal on Numerical Analysis
Publication Type :
Academic Journal
Accession number :
144779617
Full Text :
https://doi.org/10.1137/18M1186009