Inexact proximal stochastic second-order methods for nonconvex composite optimization |
| |
Authors: | Xiao Wang Hongchao Zhang |
| |
Affiliation: | 1. School of Mathematical Sciences, University of Chinese Academy of Sciences, Beijing, People's Republic of China wangxiao@ucas.ac.cn;3. Department of Mathematics, Louisiana State University, Baton Rouge, LA, USA |
| |
Abstract: | ABSTRACT In this paper, we propose a framework of Inexact Proximal Stochastic Second-order (IPSS) method for solving nonconvex optimization problems, whose objective function consists of an average of finitely many, possibly weakly, smooth functions and a convex but possibly nonsmooth function. At each iteration, IPSS inexactly solves a proximal subproblem constructed by using some positive definite matrix which could capture the second-order information of original problem. Proper tolerances are given for the subproblem solution in order to maintain global convergence and the desired overall complexity of the algorithm. Under mild conditions, we analyse the computational complexity related to the evaluations on the component gradient of the smooth function. We also investigate the number of evaluations of subgradient when using an iterative subgradient method to solve the subproblem. In addition, based on IPSS, we propose a linearly convergent algorithm under the proximal Polyak–?ojasiewicz condition. Finally, we extend the analysis to problems with weakly smooth function and obtain the computational complexity accordingly. |
| |
Keywords: | Stochastic gradient second-order approximation proximal Polyak–?ojasiewicz (PL) inequality inexact subproblem solution (weakly) smooth function variance reduction complexity nonconvex |
|
|