An Improved Reacceleration Optimization Algorithm Based on the Momentum Method for Image Recognition

被引:2
作者
Sun, Haijing [1 ]
Cai, Ying [2 ]
Tao, Ran [3 ]
Shao, Yichuan [1 ]
Xing, Lei [4 ]
Zhang, Can [2 ]
Zhao, Qian [5 ]
机构
[1] Shenyang Univ, Sch Intelligent Sci & Engn, Shenyang 110044, Peoples R China
[2] Shenyang Univ, Sch Informat Engn, Shenyang 110044, Peoples R China
[3] Shanghai Maruka Comp Informat Technol Co Ltd, Shanghai 200052, Peoples R China
[4] Univ Surrey, Sch Chem & Chem Engn, Guildford GU2 7XH, Surrey, England
[5] Shenyang Univ Technol, Sch Sci, Shenyang 110044, Peoples R China
关键词
momentum acceleration; optimization algorithm; deep learning; image recognition; gradient descent algorithm;
D O I
10.3390/math12111759
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
The optimization algorithm plays a crucial role in image recognition by neural networks. However, it is challenging to accelerate the model's convergence and maintain high precision. As a commonly used stochastic gradient descent optimization algorithm, the momentum method requires many epochs to find the optimal parameters during model training. The velocity of its gradient descent depends solely on the historical gradients and is not subject to random fluctuations. To address this issue, an optimization algorithm to enhance the gradient descent velocity, i.e., the momentum reacceleration gradient descent (MRGD), is proposed. The algorithm utilizes the point division of the current momentum and the gradient relationship, multiplying it with the gradient. It can adjust the update rate and step size of the parameters based on the gradient descent state, so as to achieve faster convergence and higher precision in training the deep learning model. The effectiveness of this method is further proven by applying the reacceleration mechanism to the Adam optimizer, resulting in the MRGDAdam algorithm. We verify both algorithms using multiple image classification datasets, and the experimental results show that the proposed optimization algorithm enables the model to achieve higher recognition accuracy over a small number of training epochs, as well as speeding up model implementation. This study provides new ideas and expansions for future optimizer research.
引用
收藏
页数:15
相关论文
共 23 条
  • [1] Dozat T., 2016, INCORPORATING NESTER
  • [2] diffGrad: An Optimization Method for Convolutional Neural Networks
    Dubey, Shiv Ram
    Chakraborty, Soumendu
    Roy, Swalpa Kumar
    Mukherjee, Snehasis
    Singh, Satish Kumar
    Chaudhuri, Bidyut Baran
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (11) : 4500 - 4511
  • [3] Guo ZS, 2024, Arxiv, DOI arXiv:2104.14840
  • [4] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [5] Heo B., 2021, P 9 INT C LEARN REPR
  • [6] Jain P., 2018, C LEARN THEOR, P545
  • [7] Kovachki N, 2021, J MACH LEARN RES, V22, P1
  • [8] Li LS, 2018, J MACH LEARN RES, V18
  • [9] [李满园 Li Manyuan], 2023, [上海理工大学学报, Journal of University of Shanghai for Science and Technology], V45, P112
  • [10] Liu L., 2020, ICLR