Stochastic Optimization for Nonconvex Problem With Inexact Hessian Matrix, Gradient, and Function

被引:2
作者
Liu, Liu [1 ,2 ]
Liu, Xuanqing [3 ]
Hsieh, Cho-Jui [4 ]
Tao, Dacheng [5 ]
机构
[1] Beihang Univ, Inst Artificial Intelligence, Beijing 100191, Peoples R China
[2] Beihang Univ, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
[3] Amazon Web Serv, Seattle, WA 98108 USA
[4] Univ Calif Los Angeles, Dept Comp Sci, Los Angeles, CA 90095 USA
[5] Univ Sydney, Sch Comp Sci, Sydney AI Ctr, Fac Engn, Sydney, NSW 2008, Australia
关键词
Optimization; Convergence; Complexity theory; Stochastic processes; Linear matrix inequalities; Approximation algorithms; Training; Adaptive regularization; stochastic optimization; trust region (TR); ALGORITHMS; REGULARIZATION; COMPLEXITY;
D O I
10.1109/TNNLS.2023.3326177
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Trust region (TR) and adaptive regularization using cubics (ARC) have proven to have some very appealing theoretical properties for nonconvex optimization by concurrently computing function value, gradient, and Hessian matrix to obtain the next search direction and the adjusted parameters. Although stochastic approximations help largely reduce the computational cost, it is challenging to theoretically guarantee the convergence rate. In this article, we explore a family of stochastic TR (STR) and stochastic ARC (SARC) methods that can simultaneously provide inexact computations of the Hessian matrix, gradient, and function values. Our algorithms require much fewer propagations overhead per iteration than TR and ARC. We prove that the iteration complexity to achieve epsilon-approximate second-order optimality is of the same order as the exact computations demonstrated in previous studies. In addition, the mild conditions on inexactness can be met by leveraging a random sampling technology in the finite-sum minimization problem. Numerical experiments with a nonconvex problem support these findings and demonstrate that, with the same or a similar number of iterations, our algorithms require less computational overhead per iteration than current secondorder methods.
引用
收藏
页码:1651 / 1663
页数:13
相关论文
共 50 条
  • [41] Inexact SARAH algorithm for stochastic optimization
    Nguyen, Lam M.
    Scheinberg, Katya
    Takac, Martin
    OPTIMIZATION METHODS & SOFTWARE, 2021, 36 (01) : 237 - 258
  • [42] A Bregman stochastic method for nonconvex nonsmooth problem beyond global Lipschitz gradient continuity
    Wang, Qingsong
    Han, Deren
    OPTIMIZATION METHODS & SOFTWARE, 2023, 38 (05) : 914 - 946
  • [43] An inexact regularized proximal Newton method for nonconvex and nonsmooth optimization
    Liu, Ruyu
    Pan, Shaohua
    Wu, Yuqia
    Yang, Xiaoqi
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2024, 88 (02) : 603 - 641
  • [44] Distributed Nonconvex Optimization: Gradient-Free Iterations and ε-Globally Optimal Solution
    He, Zhiyu
    He, Jianping
    Chen, Cailian
    Guan, Xinping
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2024, 11 (04): : 2239 - 2251
  • [45] S-DIGing: A Stochastic Gradient Tracking Algorithm for Distributed Optimization
    Li, Huaqing
    Zheng, Lifeng
    Wang, Zheng
    Yan, Yu
    Feng, Liping
    Guo, Jing
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (01): : 53 - 65
  • [46] An Inexact Augmented Lagrangian Framework for Nonconvex Optimization with Nonlinear Constraints
    Sahin, Mehmet Fatih
    Eftekhari, Armin
    Alacaoglu, Ahmet
    Latorre, Fabian
    Cevher, Volkan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [47] Inexact Block Coordinate Descent Algorithms for Nonsmooth Nonconvex Optimization
    Yang, Yang
    Pesavento, Marius
    Luo, Zhi-Quan
    Ottersten, Bjorn
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2020, 68 : 947 - 961
  • [48] STOCHASTIC ALTERNATING STRUCTURE-ADAPTED PROXIMAL GRADIENT DESCENT METHOD WITH VARIANCE REDUCTION FOR NONCONVEX NONSMOOTH OPTIMIZATION
    Jia, Zehui
    Zhang, Wenxing
    Cai, Xingju
    Han, Deren
    MATHEMATICS OF COMPUTATION, 2024, 93 (348) : 1677 - 1714
  • [49] AN INEXACT NONMONOTONE PROJECTED GRADIENT METHOD FOR CONSTRAINED MULTIOBJECTIVE OPTIMIZATION
    Zhao, Xiaopeng
    Zhang, Huijie
    Yao, Yonghong
    JOURNAL OF NONLINEAR AND VARIATIONAL ANALYSIS, 2024, 8 (04): : 517 - 531
  • [50] Stochastic Intermediate Gradient Method for Convex Problems with Stochastic Inexact Oracle
    Pavel Dvurechensky
    Alexander Gasnikov
    Journal of Optimization Theory and Applications, 2016, 171 : 121 - 145