Stochastic Optimization for Nonconvex Problem With Inexact Hessian Matrix, Gradient, and Function

被引:2
|
作者
Liu, Liu [1 ,2 ]
Liu, Xuanqing [3 ]
Hsieh, Cho-Jui [4 ]
Tao, Dacheng [5 ]
机构
[1] Beihang Univ, Inst Artificial Intelligence, Beijing 100191, Peoples R China
[2] Beihang Univ, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
[3] Amazon Web Serv, Seattle, WA 98108 USA
[4] Univ Calif Los Angeles, Dept Comp Sci, Los Angeles, CA 90095 USA
[5] Univ Sydney, Sch Comp Sci, Sydney AI Ctr, Fac Engn, Sydney, NSW 2008, Australia
关键词
Optimization; Convergence; Complexity theory; Stochastic processes; Linear matrix inequalities; Approximation algorithms; Training; Adaptive regularization; stochastic optimization; trust region (TR); ALGORITHMS; REGULARIZATION; COMPLEXITY;
D O I
10.1109/TNNLS.2023.3326177
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Trust region (TR) and adaptive regularization using cubics (ARC) have proven to have some very appealing theoretical properties for nonconvex optimization by concurrently computing function value, gradient, and Hessian matrix to obtain the next search direction and the adjusted parameters. Although stochastic approximations help largely reduce the computational cost, it is challenging to theoretically guarantee the convergence rate. In this article, we explore a family of stochastic TR (STR) and stochastic ARC (SARC) methods that can simultaneously provide inexact computations of the Hessian matrix, gradient, and function values. Our algorithms require much fewer propagations overhead per iteration than TR and ARC. We prove that the iteration complexity to achieve epsilon-approximate second-order optimality is of the same order as the exact computations demonstrated in previous studies. In addition, the mild conditions on inexactness can be met by leveraging a random sampling technology in the finite-sum minimization problem. Numerical experiments with a nonconvex problem support these findings and demonstrate that, with the same or a similar number of iterations, our algorithms require less computational overhead per iteration than current secondorder methods.
引用
收藏
页码:1651 / 1663
页数:13
相关论文
共 50 条
  • [1] Inexact Reduced Gradient Methods in Nonconvex Optimization
    Khanh, Pham Duy
    Mordukhovich, Boris S.
    Tran, Dat Ba
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2024, 203 (03) : 2138 - 2178
  • [2] A Fast Randomized Incremental Gradient Method for Decentralized Nonconvex Optimization
    Xin, Ran
    Khan, Usman A.
    Kar, Soummya
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (10) : 5150 - 5165
  • [3] Stochastic Gradient Descent for Nonconvex Learning Without Bounded Gradient Assumptions
    Lei, Yunwen
    Hu, Ting
    Li, Guiying
    Tang, Ke
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (10) : 4394 - 4400
  • [4] S-NEAR-DGD: A Flexible Distributed Stochastic Gradient Method for Inexact Communication
    Iakovidou, Charikleia
    Wei, Ermin
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (02) : 1281 - 1287
  • [5] On inexact stochastic splitting methods for a class of nonconvex composite optimization problems with relative error
    Hu, Jia
    Han, Congying
    Guo, Tiande
    Zhao, Tong
    OPTIMIZATION METHODS & SOFTWARE, 2023, 38 (01): : 1 - 33
  • [6] Stochastic inexact augmented Lagrangian method for nonconvex expectation constrained optimization
    Li, Zichong
    Chen, Pin-Yu
    Liu, Sijia
    Lu, Songtao
    Xu, Yangyang
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2024, 87 (01) : 117 - 147
  • [7] Stochastic analysis of an adaptive cubic regularization method under inexact gradient evaluations and dynamic Hessian accuracy
    Bellavia, Stefania
    Gurioli, Gianmarco
    OPTIMIZATION, 2022, 71 (01) : 227 - 261
  • [8] An inexact proximal gradient algorithm with extrapolation for a class of nonconvex nonsmooth optimization problems
    Jia, Zehui
    Wu, Zhongming
    Dong, Xiaomei
    JOURNAL OF INEQUALITIES AND APPLICATIONS, 2019, 2019 (1)
  • [9] Second-Order Guarantees of Stochastic Gradient Descent in Nonconvex Optimization
    Vlaski, Stefan
    Sayed, Ali H.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (12) : 6489 - 6504
  • [10] BLOCK STOCHASTIC GRADIENT ITERATION FOR CONVEX AND NONCONVEX OPTIMIZATION
    Xu, Yangyang
    Yin, Wotao
    SIAM JOURNAL ON OPTIMIZATION, 2015, 25 (03) : 1686 - 1716