A Stochastic Quasi-Newton Method for Large-Scale Nonconvex Optimization With Applications

被引:22
|
作者
Chen, Huiming [1 ]
Wu, Ho-Chun [1 ]
Chan, Shing-Chow [1 ]
Lam, Wong-Hing [1 ]
机构
[1] Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Peoples R China
关键词
Optimization; Stochastic processes; Convergence; Logistics; Machine learning algorithms; Linear programming; Machine learning; Damped parameter; limited memory BFGS (LBFGS); nonconjugate exponential models; nonconvex optimization; stochastic quasi-Newton (SQN) method; variational inference; VARIATIONAL INFERENCE; CLASSIFICATION; APPROXIMATION; CONVERGENCE; PROBABILITY; ALGORITHMS;
D O I
10.1109/TNNLS.2019.2957843
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Ensuring the positive definiteness and avoiding ill conditioning of the Hessian update in the stochastic BroydenFletcher-Goldfarb-Shanno (BFGS) method are significant in solving nonconvex problems. This article proposes a novel stochastic version of a damped and regularized BFGS method for addressing the above problems. While the proposed regularized strategy helps to prevent the BFGS matrix from being close to singularity, the new damped parameter further ensures the positivity of the product of correction pairs. To alleviate the computational cost of the stochastic limited memory BFGS (LBFGS) updates and to improve its robustness, the curvature information is updated using the averaged iterate at spaced intervals. The effectiveness of the proposed method is evaluated through the logistic regression and Bayesian logistic regression problems in machine learning. Numerical experiments are conducted by using both synthetic data set and several real data sets. The results show that the proposed method generally outperforms the stochastic damped LBFGS (SdLBFGS) method. In particular, for problems with small sample sizes, our method has shown superior performance and is capable of mitigating ill-conditioned problems. Furthermore, our method is more robust to the variations of the batch size and memory size than the SdLBFGS method.
引用
收藏
页码:4776 / 4790
页数:15
相关论文
共 50 条
  • [1] A STOCHASTIC QUASI-NEWTON METHOD FOR LARGE-SCALE OPTIMIZATION
    Byrd, R. H.
    Hansen, S. L.
    Nocedal, Jorge
    Singer, Y.
    SIAM JOURNAL ON OPTIMIZATION, 2016, 26 (02) : 1008 - 1031
  • [2] STOCHASTIC QUASI-NEWTON METHOD FOR NONCONVEX STOCHASTIC OPTIMIZATION
    Wang, Xiao
    Ma, Shiqian
    Goldfarb, Donald
    Liu, Wei
    SIAM JOURNAL ON OPTIMIZATION, 2017, 27 (02) : 927 - 956
  • [3] Fast large-scale optimization by unifying stochastic gradient and quasi-Newton methods
    Sohl-Dickstein, Jascha
    Poole, Ben
    Ganguli, Surya
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 604 - 612
  • [4] A Sequential Subspace Quasi-Newton Method for Large-Scale Convex Optimization
    Senov, Aleksandr
    Granichin, Oleg
    Granichina, Olga
    2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 3627 - 3632
  • [5] A stochastic extra-step quasi-Newton method for nonsmooth nonconvex optimization
    Minghan Yang
    Andre Milzarek
    Zaiwen Wen
    Tong Zhang
    Mathematical Programming, 2022, 194 : 257 - 303
  • [6] A stochastic extra-step quasi-Newton method for nonsmooth nonconvex optimization
    Yang, Minghan
    Milzarek, Andre
    Wen, Zaiwen
    Zhang, Tong
    MATHEMATICAL PROGRAMMING, 2022, 194 (1-2) : 257 - 303
  • [7] Regularization of limited memory quasi-Newton methods for large-scale nonconvex minimization
    Christian Kanzow
    Daniel Steck
    Mathematical Programming Computation, 2023, 15 : 417 - 444
  • [8] Regularization of limited memory quasi-Newton methods for large-scale nonconvex minimization
    Kanzow, Christian
    Steck, Daniel
    MATHEMATICAL PROGRAMMING COMPUTATION, 2023, 15 (03) : 417 - 444
  • [9] Preconditioned Subspace Quasi-Newton Method for Large Scale Optimization
    Sim, Hong Seng
    Leong, Wah June
    Abu Hassan, Malik
    Ismail, Fudziah
    PERTANIKA JOURNAL OF SCIENCE AND TECHNOLOGY, 2014, 22 (01): : 175 - 192
  • [10] An Overview of Stochastic Quasi-Newton Methods for Large-Scale Machine Learning
    Guo, Tian-De
    Liu, Yan
    Han, Cong-Ying
    JOURNAL OF THE OPERATIONS RESEARCH SOCIETY OF CHINA, 2023, 11 (02) : 245 - 275