Momentum Acceleration of Quasi-Newton Training for Neural Networks

被引:5
|
作者
Mahboubi, Shahrzad [1 ]
Indrapriyadarsini, S. [2 ]
Ninomiya, Hiroshi [1 ]
Asai, Hideki [2 ]
机构
[1] Shonan Inst Technol, 1-1-25 Tsujido Nishikaigan, Fujisawa, Kanagawa 2518511, Japan
[2] Shizuoka Univ, Naka Ku, 3-5-1 Johoku, Hamamatsu, Shizuoka 4328011, Japan
来源
PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT II | 2019年 / 11671卷
关键词
Neural networks; Training algorithm; Quasi-Newton method; Nesterov's accelerated quasi-Newton method; Momentum terms;
D O I
10.1007/978-3-030-29911-8_21
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes a novel acceleration technique of quasi-Newton method (QN) using momentum terms for training in neural networks. Recently, Nesterov's accelerated quasi-Newton method (NAQ) has shown that the momentum term is effective in reducing the number of iterations and in accelerating its convergence speed. However, the gradients had to calculate two times during one iteration in the NAQ training. This increased the computation time of a training loop compared with the conventional QN. In this research, an improvement to NAQ is done by approximating the Nesterov's accelerated gradient used in NAQ as a linear combination of the current and previous gradients. Then the gradient is calculated only once per iteration same as QN. The performance of the proposed algorithm is evaluated through computer simulations on a benchmark problem of the function modeling and real-world problems of the microwave circuit modeling. The results show the significant acceleration in the computation time compared with conventional training algorithms.
引用
收藏
页码:268 / 281
页数:14
相关论文
共 50 条
  • [1] Momentum acceleration of quasi-Newton based optimization technique for neural network training
    Mahboubi, Shahrzad
    Indrapriyadarsini, S.
    Ninomiya, Hiroshi
    Asai, Hideki
    IEICE NONLINEAR THEORY AND ITS APPLICATIONS, 2021, 12 (03): : 554 - 574
  • [2] A Novel quasi-Newton with Momentum Training for Microwave Circuit Models using Neural Networks
    Mahboubi, Shahrzad
    Ninomiya, Hiroshi
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON ELECTRONICS, CIRCUITS AND SYSTEMS (ICECS), 2018, : 629 - 632
  • [3] A Stochastic Momentum Accelerated Quasi-Newton Method for Neural Networks
    Indrapriyadarsini, S.
    Mahboubi, Shahrzad
    Ninomiya, Hiroshi
    Kamio, Takeshi
    Asai, Hideki
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 12973 - 12974
  • [4] Modified quasi-Newton methods for training neural networks
    Robitaille, B
    Marcos, B
    Veillette, M
    Payre, G
    COMPUTERS & CHEMICAL ENGINEERING, 1996, 20 (09) : 1133 - 1140
  • [5] Practical Quasi-Newton Methods for Training Deep Neural Networks
    Goldfarb, Donald
    Ren, Yi
    Bahamou, Achraf
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [6] On the study of the memory-less quasi-Newton method with momentum term for neural network training
    Mahboubi, Shahrzad
    Yamatomi, Ryo
    Indrapriyadarsini, S.
    Ninomiya, Hiroshi
    Asai, Hideki
    IEICE NONLINEAR THEORY AND ITS APPLICATIONS, 2022, 13 (02): : 271 - 276
  • [7] A QUASI-NEWTON ACCELERATION OF THE EM ALGORITHM
    LANGE, K
    STATISTICA SINICA, 1995, 5 (01) : 1 - 18
  • [8] Deep Neural Networks Training by Stochastic Quasi-Newton Trust-Region Methods
    Yousefi, Mahsa
    Martinez, Angeles
    ALGORITHMS, 2023, 16 (10)
  • [9] Training the random neural network using quasi-Newton methods
    Likas, A
    Stafylopatis, A
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2000, 126 (02) : 331 - 339
  • [10] Quasi-Newton Learning Methods for Complex-Valued Neural Networks
    Popa, Calin-Adrian
    2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,