Fast stochastic second-order method logarithmic in condition number

被引:1
|
作者
Ye, Haishan [1 ]
Xie, Guangzeng [2 ]
Luo, Luo [1 ]
Zhang, Zhihua [2 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Comp Sci & Engn, 800 Dong Chuan Rd, Shanghai 200240, Peoples R China
[2] Peking Univ, Sch Math Sci, Beijing 100871, Peoples R China
基金
中国国家自然科学基金;
关键词
38;
D O I
10.1016/j.patcog.2018.11.031
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Optimization is an important issue in machine learning because many machine learning models are reformulated as optimization problems. Different kinds of machine learning algorithms mainly focus on minimizing their empirical loss like deep learning, logistic regression, and support vector machine. Because data is explosively growing, it is challenging to deal with a large-scale optimization problem. Recently, stochastic second-order methods have emerged to attract much attention due to their efficiency in each iteration. These methods show good performance on training machine learning algorithms like logistic regression and support vector machine. However, the computational complexity of existing stochastic second-order methods heavily depends on the condition number of the Hessian. In this paper, we propose a new Newton-like method called Preconditioned Newton Conjugate Gradient with Sketched Hessian (PNCG). The runtime complexity of PNCG is at most logarithmic in the condition number of the Hessian. PNCG exhibits advantages over existing subsampled Newton methods especially when the Hessian matrix in question is ill-conditioned. We also show that our method has good performance on training machine learning algorithm empirically. The results show consistent improvements in computational efficiency. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:629 / 642
页数:14
相关论文
共 50 条
  • [1] Fast Second-Order Stochastic Backpropagation for Variational Inference
    Fan, Kai
    Wang, Ziteng
    Beck, Jeffrey
    Kwok, James T.
    Heller, Katherine
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [2] SECOND-ORDER FAST-SLOW STOCHASTIC SYSTEMS
    Nguyen, Nhu N.
    Yin, George
    SIAM JOURNAL ON MATHEMATICAL ANALYSIS, 2024, 56 (04) : 5175 - 5208
  • [3] How Tight is the Necessary Condition for the Second-order Stochastic Dominance?
    Kopa, Milos
    38TH INTERNATIONAL CONFERENCE ON MATHEMATICAL METHODS IN ECONOMICS (MME 2020), 2020, : 301 - 306
  • [4] An Accelerated Second-Order Method for Distributed Stochastic Optimization
    Agafonov, Artem
    Dvurechensky, Pavel
    Scutari, Gesualdo
    Gasnikov, Alexander
    Kamzolov, Dmitry
    Lukashevich, Aleksandr
    Daneshmand, Amir
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 2407 - 2413
  • [5] A Second-Order Method for Stochastic Bandit Convex Optimisation
    Lattimore, Tor
    Gyorgy, Andras
    THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195
  • [6] A fully stochastic second-order trust region method
    Curtis, Frank E.
    Shi, Rui
    Optimization Methods and Software, 2022, 37 (03): : 844 - 877
  • [7] A fully stochastic second-order trust region method
    Curtis, Frank E.
    Shi, Rui
    OPTIMIZATION METHODS & SOFTWARE, 2022, 37 (03): : 844 - 877
  • [8] A Stochastic Second-Order Proximal Method for Distributed Optimization
    Qiu, Chenyang
    Zhu, Shanying
    Ou, Zichong
    Lu, Jie
    IEEE CONTROL SYSTEMS LETTERS, 2023, 7 : 1405 - 1410
  • [9] A Second-Order Evolution Equation and Logarithmic Operators
    F. D. M. Bezerra
    Bulletin of the Brazilian Mathematical Society, New Series, 2022, 53 : 571 - 593
  • [10] Number as a second-order concept
    Damerow, P
    SCIENCE IN CONTEXT, 1996, 9 (02) : 139 - 149