首页 | 本学科首页   官方微博 | 高级检索  
     


Inexact proximal stochastic second-order methods for nonconvex composite optimization
Authors:Xiao Wang  Hongchao Zhang
Affiliation:1. School of Mathematical Sciences, University of Chinese Academy of Sciences, Beijing, People's Republic of China wangxiao@ucas.ac.cn;3. Department of Mathematics, Louisiana State University, Baton Rouge, LA, USA
Abstract:ABSTRACT

In this paper, we propose a framework of Inexact Proximal Stochastic Second-order (IPSS) method for solving nonconvex optimization problems, whose objective function consists of an average of finitely many, possibly weakly, smooth functions and a convex but possibly nonsmooth function. At each iteration, IPSS inexactly solves a proximal subproblem constructed by using some positive definite matrix which could capture the second-order information of original problem. Proper tolerances are given for the subproblem solution in order to maintain global convergence and the desired overall complexity of the algorithm. Under mild conditions, we analyse the computational complexity related to the evaluations on the component gradient of the smooth function. We also investigate the number of evaluations of subgradient when using an iterative subgradient method to solve the subproblem. In addition, based on IPSS, we propose a linearly convergent algorithm under the proximal Polyak–?ojasiewicz condition. Finally, we extend the analysis to problems with weakly smooth function and obtain the computational complexity accordingly.
Keywords:Stochastic gradient  second-order approximation  proximal Polyak–?ojasiewicz (PL) inequality  inexact subproblem solution  (weakly) smooth function  variance reduction  complexity  nonconvex
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号