24 results on '"Nesterov’s acceleration"'
Search Results
2. Accelerated Deep Nonlinear Dictionary Learning
- Author
-
Tan, Benying, Lin, Jie, Qin, Yang, Ding, Shuxue, Li, Yujie, Goos, Gerhard, Series Editor, Hartmanis, Juris, Founding Editor, Bertino, Elisa, Editorial Board Member, Gao, Wen, Editorial Board Member, Steffen, Bernhard, Editorial Board Member, Yung, Moti, Editorial Board Member, Cho, Minsu, editor, Laptev, Ivan, editor, Tran, Du, editor, Yao, Angela, editor, and Zha, Hongbin, editor
- Published
- 2025
- Full Text
- View/download PDF
3. Stagewise Accelerated Stochastic Gradient Methods for Nonconvex Optimization.
- Author
-
Jia, Cui and Cui, Zhuoxu
- Subjects
- *
ARTIFICIAL neural networks , *DEEP learning , *MACHINE learning - Abstract
For large-scale optimization that covers a wide range of optimization problems encountered frequently in machine learning and deep neural networks, stochastic optimization has become one of the most used methods thanks to its low computational complexity. In machine learning and deep learning problems, nonconvex problems are common, while convex problems are rare. How to find the global minimum for nonconvex optimization and reduce the computational complexity are challenges. Inspired by the phenomenon that the stagewise stepsize tuning strategy can empirically improve the convergence speed in deep neural networks, we incorporate the stagewise stepsize tuning strategy into the iterative framework of Nesterov's acceleration- and variance reduction-based methods to reduce the computational complexity, i.e., the stagewise stepsize tuning strategy is incorporated into randomized stochastic accelerated gradient and stochastic variance-reduced gradient. The proposed methods are theoretically derived to reduce the complexity of the nonconvex and convex problems and improve the convergence rate of the frameworks, which have the complexity O (1 / μ ϵ) and O (1 / μ ϵ) , respectively, where μ is the PL modulus and L is the Lipschitz constant. In the end, numerical experiments on large benchmark datasets validate well the competitiveness of the proposed methods. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
4. Extragradient-type methods with O1/k last-iterate convergence rates for co-hypomonotone inclusions.
- Author
-
Tran-Dinh, Quoc
- Subjects
SET-valued maps ,RESOLVENTS (Mathematics) - Abstract
We develop two "Nesterov's accelerated" variants of the well-known extragradient method to approximate a solution of a co-hypomonotone inclusion constituted by the sum of two operators, where one is Lipschitz continuous and the other is possibly multivalued. The first scheme can be viewed as an accelerated variant of Tseng's forward-backward-forward splitting (FBFS) method, while the second one is a Nesterov's accelerated variant of the "past" FBFS scheme, which requires only one evaluation of the Lipschitz operator and one resolvent of the multivalued mapping. Under appropriate conditions on the parameters, we theoretically prove that both algorithms achieve O 1 / k last-iterate convergence rates on the residual norm, where k is the iteration counter. Our results can be viewed as alternatives of a recent class of Halpern-type methods for root-finding problems. For comparison, we also provide a new convergence analysis of the two recent extra-anchored gradient-type methods for solving co-hypomonotone inclusions. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
5. Event-Triggered Acceleration Algorithms for Distributed Stochastic Optimization
- Author
-
Lü, Qingguo, Liao, Xiaofeng, Li, Huaqing, Deng, Shaojiang, Gao, Shanfu, Shen, Xuemin Sherman, Series Editor, Lü, Qingguo, Liao, Xiaofeng, Li, Huaqing, Deng, Shaojiang, and Gao, Shanfu
- Published
- 2023
- Full Text
- View/download PDF
6. Accelerating regularized tensor decomposition using the alternating direction method of multipliers with multiple Nesterov's extrapolations.
- Author
-
Wang, Deqing and Hu, Guoqiang
- Subjects
- *
MULTIPLIERS (Mathematical analysis) , *EXTRAPOLATION , *SIGNAL processing , *ALGORITHMS - Abstract
Tensor decomposition is an essential tool for multiway signal processing. At present, large-scale high-order tensor data require fast and efficient decomposing algorithms. In this paper, we propose accelerated regularized tensor decomposition algorithms using the alternating direction method of multipliers with multiple Nesterov's extrapolations in the block coordinate descent framework. We implement the acceleration in three cases: only in the inner loop, only in the outer loop, and in both the inner and outer loops. Adaptive safeguard strategies are developed following the acceleration to guarantee monotonic convergence. Afterwards, we utilize the proposed algorithms to accelerate two types of conventional decomposition: nonnegative CANDECOMP/PARAFAC (NCP) and sparse CANDECOMP/PARAFAC (SCP). The experimental results on synthetic and real-world tensors demonstrate that the proposed algorithms achieve significant acceleration effects and outperform state-of-the-art algorithms. The accelerated algorithm with extrapolations in both the inner and outer loops has the fastest convergence speed and takes almost one-third of the running time of typical algorithms. • Accelerate regularized tensor decomposition using ADMM with multiple extrapolations. • Extrapolations are utilized in three cases: inner loop, outer loop, and both loops. • Adaptive safeguard strategies are developed to guarantee monotonic convergence. • The accelerated algorithms converge in almost one-third of conventional running time. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. BREGMAN PROXIMAL POINT ALGORITHM REVISITED: A NEW INEXACT VERSION AND ITS INERTIAL VARIANT.
- Author
-
LEI YANG and KIM-CHUAN TOH
- Subjects
- *
COMMUNITIES , *KERNEL functions , *PROBLEM solving - Abstract
We study a general convex optimization problem, which covers various classic problems in different areas and particularly includes many optimal transport related problems arising in recent years. To solve this problem, we revisit the classic Bregman proximal point algorithm (BPPA) and introduce a new inexact stopping condition for solving the subproblems, which can circumvent the underlying feasibility difficulty often appearing in existing inexact conditions when the problem has a complex feasible set. Our inexact condition also covers several existing inexact conditions as special cases and hence makes our inexact BPPA (iBPPA) more flexible to fit different scenarios in practice. As an application to the standard optimal transport (OT) problem, our iBPPA with the entropic proximal term can bypass some numerical instability issues that usually plague the popular Sinkhorn's algorithm in the OT community, since our iBPPA does not require the proximal parameter to be very small for obtaining an accurate approximate solution. The iteration complexity of O(1/k) and the convergence of the sequence are also established for our iBPPA under some mild conditions. Moreover, inspired by Nesterov's acceleration technique, we develop an inertial variant of our iBPPA, denoted by V-iBPPA, and establish the iteration complexity of O(1/kλ), where λ ≥ 1 is a quadrangle scaling exponent of the kernel function. In particular, when the proximal parameter is a constant and the kernel function is strongly convex with Lipschitz continuous gradient (hence λ = 2), our V-iBPPA achieves a faster rate of O(1/k²) just as existing accelerated inexact proximal point algorithms. Some preliminary numerical experiments for solving the standard OT problem are conducted to show the convergence behaviors of our iBPPA and V-iBPPA under different inexactness settings. The experiments also empirically verify the potential of our V-iBPPA for improving the convergence speed. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
8. Sampling Kaczmarz-Motzkin method for linear feasibility problems: generalization and acceleration.
- Author
-
Morshed, Md Sarowar, Islam, Md Saiful, and Noor-E-Alam, Md.
- Subjects
- *
INTERIOR-point methods , *SUPPORT vector machines , *SAMPLING methods , *GENERALIZATION , *NAIVE Bayes classification , *LINEAR equations - Abstract
Randomized Kaczmarz, Motzkin Method and Sampling Kaczmarz Motzkin (SKM) algorithms are commonly used iterative techniques for solving a system of linear inequalities (i.e., A x ≤ b ). As linear systems of equations represent a modeling paradigm for solving many optimization problems, these randomized and iterative techniques are gaining popularity among researchers in different domains. In this work, we propose a Generalized Sampling Kaczmarz Motzkin (GSKM) method that unifies the iterative methods into a single framework. In addition to the general framework, we propose a Nesterov-type acceleration scheme in the SKM method called Probably Accelerated Sampling Kaczmarz Motzkin (PASKM). We prove the convergence theorems for both GSKM and PASKM algorithms in the L 2 norm perspective with respect to the proposed sampling distribution. Furthermore, we prove sub-linear convergence for the Cesaro average of iterates for the proposed GSKM and PASKM algorithms. From the convergence theorem of the GSKM algorithm, we find the convergence results of several well-known algorithms like the Kaczmarz method, Motzkin method and SKM algorithm. We perform thorough numerical experiments using both randomly generated and real-world (classification with support vector machine and Netlib LP) test instances to demonstrate the efficiency of the proposed methods. We compare the proposed algorithms with SKM, Interior Point Method and Active Set Method in terms of computation time and solution quality. In the majority of the problem instances, the proposed generalized and accelerated algorithms significantly outperform the state-of-the-art methods. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
9. An Edge Storage Acceleration Service for Collaborative Mobile Devices.
- Author
-
Gao, Xiong, Bao, Weidong, Zhu, Xiaomin, Wu, Guanlin, and Liu, Ling
- Abstract
Fueled by the advances in the Internet of Things, and the growing capacity of smart mobile devices at the edge of the Internet, we have witnessed a growing trend in research and development for edge computing and edge storage, which extends the abilities of single mobile device on the edge through on-demand collaboration among multiple geographically distributed mobile devices. In this article, we address several technical challenges that are unique for collaborative storage at the edge due to the unique characteristics of mobile devices. First, we formalize the collaborative storage problem as an optimization problem. Second, we design an Acceleration Algorithm for Collaborative Storage, called A2CS, based on the architecture of Alternating Direction Method of Multipliers (ADMM). Specifically, we use the Nesterov’s Acceleration strategy and the step size rules in the process of updating variables and determining the optimal speed of convergence. We develop a novel collaborative storage policy in order to guide the whole lifecycle of collaborative storage. Finally, we conduct a series of experiments for acceleration performance analysis and validation. We show that A2CS delivers a better convergence performance with different step size rules, compared with two existing approaches: the ADMM baseline and the ADMM-OR (ADMM with Over-Relaxation), achieving the acceleration percentage by at least 25.33 percent and at most 64.01 percent. In addition, by conducting the utility performance comparison analysis with the existing Average Distribution Strategy (ADS) and the existing Distance Preferred Distribution Strategy (DPDS), we show the advantage of A2CS over both ADS and DPDS with respect to the total utility and energy consumption. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
10. Multi-Time Scale Smoothed Functional With Nesterov’s Acceleration
- Author
-
Abhinav Sharma, K. Lakshmanan, Ruchir Gupta, and Atul Gupta
- Subjects
Multi-Stage queueing networks ,Nesterov’s acceleration ,simulation ,smoothed functional algorithm ,stochastic approximation algorithms ,stochastic optimization ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Smoothed functional (SF) algorithm estimates the gradient of the stochastic optimization problem by convolution with a smoothening kernel. This process helps the algorithm to converge to a global minimum or a point close to it. We study a two-time scale SF based gradient search algorithm with Nesterov’s acceleration for stochastic optimization problems. The main contribution of our work is to prove the convergence of this algorithm using the stochastic approximation theory. We propose a novel Lyapunov function to show the associated second-order ordinary differential equations’ (o.d.e.) stability for a non-autonomous system. We compare our algorithm with other smoothed functional algorithms such as Quasi-Newton SF, Gradient SF and Jacobi Variant of Newton SF on two different optimization problems: first, on a simple stochastic function minimization problem, and second, on the problem of optimal routing in a queueing network. Additionally, we compared the algorithms on real weather data in a weather prediction task. Experimental results show that our algorithm performs significantly better than these baseline algorithms.
- Published
- 2021
- Full Text
- View/download PDF
11. Revisiting Randomized Gossip Algorithms: General Framework, Convergence Rates and Novel Block and Accelerated Protocols.
- Author
-
Loizou, Nicolas and Richtarik, Peter
- Subjects
- *
GOSSIP , *ALGORITHMS , *DISTRIBUTED algorithms , *BLOCK designs , *LINEAR systems - Abstract
In this work we present a new framework for the analysis and design of randomized gossip algorithms for solving the average consensus problem. We show how classical randomized iterative methods for solving linear systems can be interpreted as gossip algorithms when applied to special systems encoding the underlying network and explain in detail their decentralized nature. Our general framework recovers a comprehensive array of well-known gossip algorithms as special cases, including the pairwise randomized gossip algorithm and path averaging gossip, and allows for the development of provably faster variants. The flexibility of the new approach enables the design of a number of new specific gossip methods. For instance, we propose and analyze novel block and the first provably accelerated randomized gossip protocols, and dual randomized gossip algorithms. From a numerical analysis viewpoint, our work is the first that explores in depth the decentralized nature of randomized iterative methods for linear systems and proposes them as methods for solving the average consensus problem. We evaluate the performance of the proposed gossip protocols by performing extensive experimental testing on typical wireless network topologies. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
12. Application of a class of iterative algorithms and their accelerations to Jacobian-based linearized EIT image reconstruction.
- Author
-
Wang, Jing and Han, Bo
- Subjects
- *
IMAGE reconstruction algorithms , *IMAGE reconstruction , *ELECTRICAL impedance tomography , *ALGORITHMS - Abstract
This work is concerned with the image reconstruction of the Jacobian-based linearized EIT problem. Based on the homotopy perturbation technology, we first propose a novel class of iteration schemes with different orders of approximation truncation (named as HPI for short), which contains Landweber-type iteration method. Afterwards, nonsmooth priors such as ℓ 1 -norm or total variation penalty are introduced with the proposed HPI method to improve the imaging quality. However, it is known that Landweber-type iteration method is the widely used imaging algorithm, but in its basic form, the imaging speed is slow and the imaging accuracy is low. Furthermore, Nesterov's strategy is conducted to accelerate the proposed approaches. Numerical simulations with synthetic dada are performed to validate that our proposed approaches have robustness to noise and can indeed improve the image resolution and the imaging speed. Especially, numerical results explicitly show that Nesterov's acceleration versions make significantly remarkable acceleration effects. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
13. Accelerated Proximal Subsampled Newton Method.
- Author
-
Ye, Haishan, Luo, Luo, and Zhang, Zhihua
- Subjects
- *
NEWTON-Raphson method , *PROBLEM solving , *ALGORITHMS , *MACHINE learning , *EMPIRICAL research - Abstract
Composite function optimization problem often arises in machine learning known as regularized empirical minimization. We introduce the acceleration technique to the Newton-type proximal method and propose a novel algorithm called accelerated proximal subsampled Newton method (APSSN). APSSN only subsamples a small subset of samples to construct an approximate Hessian that achieves computational efficiency. At the same time, APSSN still keeps a fast convergence rate. Furthermore, we obtain the scaled proximal mapping by solving its dual problem using the semismooth Newton method instead of resorting to the first-order methods. Due to our sampling strategy and the fast convergence rate of the semismooth Newton method, we can get the scaled proximal mapping efficiently. Both our theoretical analysis and empirical study show that APSSN is an effective and computationally efficient algorithm for composite function optimization problems. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
14. Accelerated sampling Kaczmarz Motzkin algorithm for the linear feasibility problem.
- Author
-
Morshed, Md Sarowar, Islam, Md Saiful, and Noor-E-Alam, Md.
- Subjects
INTERIOR-point methods ,RELAXATION methods (Mathematics) ,ALGORITHMS ,LINEAR systems - Abstract
The Sampling Kaczmarz Motzkin (SKM) algorithm is a generalized method for solving large-scale linear systems of inequalities. Having its root in the relaxation method of Agmon, Schoenberg, and Motzkin and the randomized Kaczmarz method, SKM outperforms the state-of-the-art methods in solving large-scale Linear Feasibility (LF) problems. Motivated by SKM's success, in this work, we propose an Accelerated Sampling Kaczmarz Motzkin (ASKM) algorithm which achieves better convergence compared to the standard SKM algorithm on ill-conditioned problems. We provide a thorough convergence analysis for the proposed accelerated algorithm and validate the results with various numerical experiments. We compare the performance and effectiveness of ASKM algorithm with SKM, Interior Point Method (IPM) and Active Set Method (ASM) on randomly generated instances as well as Netlib LPs. In most of the test instances, the proposed ASKM algorithm outperforms the other state-of-the-art methods. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
15. Nesterov's Acceleration for Approximate Newton.
- Author
-
Haishan Ye, Luo Luo, and Zhihua Zhang
- Subjects
- *
MACHINE learning , *DILEMMA - Abstract
Optimization plays a key role in machine learning. Recently, stochastic second-order methods have attracted considerable attention because of their low computational cost in each iteration. However, these methods might suffer from poor performance when the Hessian is hard to be approximate well in a computation-efficient way. To overcome this dilemma, we resort to Nesterov's acceleration to improve the convergence performance of these second-order methods and propose accelerated approximate Newton. We give the theoretical convergence analysis of accelerated approximate Newton and show that Nesterov's acceleration can improve the convergence rate. Accordingly, we propose an accelerated regularized sub-sampled Newton (ARSSN) which performs much better than the conventional regularized sub-sampled Newton empirically and theoretically. Moreover, we show that ARSSN has better performance than classical first-order methods empirically. [ABSTRACT FROM AUTHOR]
- Published
- 2020
16. New analysis of linear convergence of gradient-type methods via unifying error bound conditions.
- Author
-
Zhang, Hui
- Subjects
- *
FORWARD-backward algorithm , *CONJUGATE gradient methods , *CONVEXITY spaces , *ALGORITHMS - Abstract
This paper reveals that a common and central role, played in many error bound (EB) conditions and a variety of gradient-type methods, is a residual measure operator. On one hand, by linking this operator with other optimality measures, we define a group of abstract EB conditions, and then analyze the interplay between them; on the other hand, by using this operator as an ascent direction, we propose an abstract gradient-type method, and then derive EB conditions that are necessary and sufficient for its linear convergence. The former provides a unified framework that not only allows us to find new connections between many existing EB conditions, but also paves a way to construct new ones. The latter allows us to claim the weakest conditions guaranteeing linear convergence for a number of fundamental algorithms, including the gradient method, the proximal point algorithm, and the forward–backward splitting algorithm. In addition, we show linear convergence for the proximal alternating linearized minimization algorithm under a group of equivalent EB conditions, which are strictly weaker than the traditional strongly convex condition. Moreover, by defining a new EB condition, we show Q-linear convergence of Nesterov's accelerated forward–backward algorithm without strong convexity. Finally, we verify EB conditions for a class of dual objective functions. [ABSTRACT FROM AUTHOR]
- Published
- 2020
- Full Text
- View/download PDF
17. A Multicategory Kernel Distance Weighted Discrimination Method for Multiclass Classification.
- Author
-
Wang, Boxiang and Zou, Hui
- Subjects
- *
HILBERT space , *SPEECH perception , *REMOTE-sensing images , *CLASSIFICATION - Abstract
Distance weighted discrimination (DWD) is an interesting large margin classifier that has been shown to enjoy nice properties and empirical successes. The original DWD only handles binary classification with a linear classification boundary. Multiclass classification problems naturally appear in various fields, such as speech recognition, satellite imagery classification, and self-driving vehicles, to name a few. For such complex classification problems, it is desirable to have a flexible multicategory kernel extension of the binary DWD when the optimal decision boundary is highly nonlinear. To this end, we propose a new multicategory kernel DWD, that is, defined as a margin-vector optimization problem in a reproducing kernel Hilbert space. This formulation is shown to enjoy Fisher consistency. We develop an accelerated projected gradient descent algorithm to fit the multicategory kernel DWD. Simulations and benchmark data applications are used to demonstrate the highly competitive performance of our method, as compared with some popular state-of-the-art multiclass classifiers. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
18. Accelerated gradient boosting.
- Author
-
Biau, G., Cadre, B., and Rouvière, L.
- Subjects
DECISION trees ,PROBLEM solving ,BOOSTING algorithms ,MATHEMATICAL optimization ,DATA - Abstract
Gradient tree boosting is a prediction algorithm that sequentially produces a model in the form of linear combinations of decision trees, by solving an infinite-dimensional optimization problem. We combine gradient boosting and Nesterov's accelerated descent to design a new algorithm, which we call AGB (for Accelerated Gradient Boosting). Substantial numerical evidence is provided on both synthetic and real-life data sets to assess the excellent performance of the method in a large variety of prediction problems. It is empirically shown that AGB is less sensitive to the shrinkage parameter and outputs predictors that are considerably more sparse in the number of trees, while retaining the exceptional performance of gradient boosting. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
19. An iterative reconstruction algorithm for unsupervised PET image.
- Author
-
Wang S, Liu B, Xie F, and Chai L
- Subjects
- Humans, Tomography, X-Ray Computed, Algorithms, Neural Networks, Computer, Phantoms, Imaging, Image Processing, Computer-Assisted methods, Positron-Emission Tomography methods
- Abstract
Objective. In recent years, convolutional neural networks (CNNs) have shown great potential in positron emission tomography (PET) image reconstruction. However, most of them rely on many low-quality and high-quality reference PET image pairs for training, which are not always feasible in clinical practice. On the other hand, many works improve the quality of PET image reconstruction by adding explicit regularization or optimizing the network structure, which may lead to complex optimization problems. Approach. In this paper, we develop a novel iterative reconstruction algorithm by integrating the deep image prior (DIP) framework, which only needs the prior information (e.g. MRI) and sinogram data of patients. To be specific, we construct the objective function as a constrained optimization problem and utilize the existing PET image reconstruction packages to streamline calculations. Moreover, to further improve both the reconstruction quality and speed, we introduce the Nesterov's acceleration part and the restart mechanism in each iteration. Main results. 2D experiments on PET data sets based on computer simulations and real patients demonstrate that our proposed algorithm can outperform existing MLEM-GF, KEM and DIPRecon methods. Significance. Unlike traditional CNN methods, the proposed algorithm does not rely on large data sets, but only leverages inter-patient information. Furthermore, we enhance reconstruction performance by optimizing the iterative algorithm. Notably, the proposed method does not require much modification of the basic algorithm, allowing for easy integration into standard implementations., (© 2024 Institute of Physics and Engineering in Medicine.)
- Published
- 2024
- Full Text
- View/download PDF
20. An Accelerated Linearly Convergent Stochastic L-BFGS Algorithm.
- Author
-
Chang, Daqing, Sun, Shiliang, and Zhang, Changshui
- Subjects
- *
ALGORITHMS , *MACHINE learning , *INTERIOR-point methods - Abstract
The limited memory version of the Broyden–Fletcher–Goldfarb–Shanno (L-BFGS) algorithm is the most popular quasi-Newton algorithm in machine learning and optimization. Recently, it was shown that the stochastic L-BFGS (sL-BFGS) algorithm with the variance-reduced stochastic gradient converges linearly. In this paper, we propose a new sL-BFGS algorithm by importing a proper momentum. We prove an accelerated linear convergence rate under mild conditions. The experimental results on different data sets also verify this acceleration advantage. [ABSTRACT FROM AUTHOR]
- Published
- 2019
- Full Text
- View/download PDF
21. Three-Dimensional Data Compression and Fast High-Quality Reconstruction for Phased Array Weather Radar
- Author
-
Tomoo Ushio, Ryosuke Kawami, Daichi Kitahara, Hiroshi Kikuchi, Eiichi Yoshikawa, and Akira Hirabayashi
- Subjects
convex optimization ,Computer science ,Phased array ,Acoustics ,Phased array weather radar ,law.invention ,Quality (physics) ,Compressed sensing ,law ,Compression (functional analysis) ,Convex optimization ,Three dimensional data ,Weather radar ,Electrical and Electronic Engineering ,Nesterov's acceleration ,data compression ,compressed sensing ,Data compression - Abstract
資料番号: PA2020054000
- Published
- 2020
- Full Text
- View/download PDF
22. Fast primal–dual algorithm via dynamical system for a linearly constrained convex optimization problem.
- Author
-
He, Xin, Hu, Rong, and Fang, Ya-Ping
- Subjects
- *
ALGORITHMS , *CONSTRAINED optimization , *DYNAMICAL systems , *LINEAR dynamical systems - Abstract
By time discretization of a second-order primal–dual dynamical system with damping α / t where an inertial construction in the sense of Nesterov is needed only for the primal variable, we propose a fast primal–dual algorithm for a linear equality constrained convex optimization problem. Under a suitable scaling condition, we show that the proposed algorithm enjoys a fast convergence rate for the objective residual and the feasibility violation, and the decay rate can reach O (1 / k α − 1) at the most. We also study convergence properties of the corresponding primal–dual dynamical system to better understand the acceleration scheme. Finally, we report numerical experiments to demonstrate the effectiveness of the proposed algorithm. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
23. Nesterov's Acceleration for Approximate Newton
- Author
-
Ye, Haishan, Luo, Luo, Zhang, Zhihua, Ye, Haishan, Luo, Luo, and Zhang, Zhihua
- Abstract
Optimization plays a key role in machine learning. Recently, stochastic second-order methods have attracted considerable attention because of their low computational cost in each iteration. However, these methods might suffer from poor performance when the Hessian is hard to be approximate well in a computation-efficient way. To overcome this dilemma, we resort to Nesterov's acceleration to improve the convergence performance of these second-order methods and propose accelerated approximate Newton. We give the theoretical convergence analysis of accelerated approximate Newton and show that Nesterov's acceleration can improve the convergence rate. Accordingly, we propose an accelerated regularized sub-sampled Newton (ARS SN) which performs much better than the conventional regularized sub-sampled Newton empirically and theoretically. Moreover, we show that ARS SN has better performance than classical first-order methods empirically.
- Published
- 2020
24. An accelerated IRNN-Iteratively Reweighted Nuclear Norm algorithm for nonconvex nonsmooth low-rank minimization problems.
- Author
-
Phan, Duy Nhat and Nguyen, Thuy Ngoc
- Subjects
- *
NONSMOOTH optimization , *RECOMMENDER systems , *ALGORITHMS , *SMOOTHNESS of functions , *MACHINE learning - Abstract
Low-rank minimization problems arise in numerous important applications such as recommendation systems, machine learning, network analysis, and so on. The problems however typically consist of minimizing a sum of a smooth function and nonconvex nonsmooth composite functions, which solving them remains a big challenge. In this paper, we take inspiration from the Nesterov's acceleration technique to accelerate an iteratively reweighted nuclear norm algorithm for the considered problems ensuring that every limit point is a critical point. Our algorithm iteratively computes the proximal operator of a reweighted nuclear norm which has a closed-form solution by performing the SVD of a smaller matrix instead of the full SVD. This distinguishes our work from recent accelerated proximal gradient algorithms that require an expensive computation of the proximal operator of nonconvex nonsmooth composite functions. We also investigate the convergence rate with the Kurdyka–Łojasiewicz assumption. Numerical experiments are performed to demonstrate the efficiency of our algorithm and its superiority over well-known methods. [ABSTRACT FROM AUTHOR]
- Published
- 2021
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.