Improving the Accuracy of Neural Network Pattern Recognition by Fractional Gradient Descent

被引:1
|
作者
Abdulkadirov, Ruslan I. [1 ]
Lyakhov, Pavel A. [1 ]
Baboshina, Valentina A. [1 ]
Nagornov, Nikolay N. [2 ]
机构
[1] North Caucasus Fed Univ, North Caucasus Ctr Math Res, Stavropol 355017, Russia
[2] North Caucasus Fed Univ, Dept Math Modeling, Stavropol 355017, Russia
来源
IEEE ACCESS | 2024年 / 12卷
基金
俄罗斯科学基金会;
关键词
Neural networks; Optimization; Pattern recognition; Accuracy; Convergence; Training; Transformers; Heuristic algorithms; Stability analysis; Multilayer perceptrons; Convolutional neural networks; fractional derivatives of Riemann-Liouville; Caputo; Grunwald-Letnikov; multilayer perceptron; optimization algorithms; stochastic gradient descent; OPTIMIZATION;
D O I
10.1109/ACCESS.2024.3491614
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper we propose the fractional gradient descent for increasing the training and work of modern neural networks. This optimizer searches the global minimum of the loss function considering the fractional gradient directions achieved by Riemann-Liouville, Caputo, and Grunwald-Letnikov derivatives. The adjusting of size and direction of the fractional gradient, supported by momentum and Nesterov condition, let the proposed optimizer descend into the global minimum of loss functions of neural networks. Utilizing the proposed optimization algorithm in a linear neural network and a visual transformer lets us attain higher accuracy, precision, recall, Macro F1 score by 1.8-4 percentage points than known analogs than state-of-the-art methods in solving pattern recognition problems on images from MNIST and CIFAR10 datasets. Further research of fractional calculus in modern neural network methodology can improve the quality of solving various challenges such as pattern recognition, time series forecasting, moving object detection, and data generation.
引用
收藏
页码:168428 / 168444
页数:17
相关论文
共 50 条
  • [1] Fault diagnosis of rolling bearing based on BP neural network with fractional order gradient descent
    Jiao, Rui
    Li, Sai
    Ding, Zhixia
    Yang, Le
    Wang, Guan
    JOURNAL OF VIBRATION AND CONTROL, 2024, 30 (9-10) : 2139 - 2153
  • [2] Training Neural Networks by Time-Fractional Gradient Descent
    Xie, Jingyi
    Li, Sirui
    AXIOMS, 2022, 11 (10)
  • [3] TE/TM Pattern Recognition Based on Convolutional Neural Network
    Chu, Mingxin
    Yu, Peng
    Che, Ping
    Guan, Xiaofei
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2025, 22
  • [4] Convergence of Stochastic Gradient Descent in Deep Neural Network
    Bai-cun Zhou
    Cong-ying Han
    Tian-de Guo
    Acta Mathematicae Applicatae Sinica, English Series, 2021, 37 : 126 - 136
  • [5] Convergence of Stochastic Gradient Descent in Deep Neural Network
    Zhou, Bai-cun
    Han, Cong-ying
    Guo, Tian-de
    ACTA MATHEMATICAE APPLICATAE SINICA-ENGLISH SERIES, 2021, 37 (01): : 126 - 136
  • [6] Study on fast speed fractional order gradient descent method and its application in neural networks
    Wang, Yong
    He, Yuli
    Zhu, Zhiguang
    NEUROCOMPUTING, 2022, 489 : 366 - 376
  • [7] Feedback neural network for pattern recognition
    Salih, I
    Smith, SH
    APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN IMAGE PROCESSING IV, 1999, 3647 : 194 - 201
  • [8] Stochastic Gradient Descent Method of Convolutional Neural Network Using Fractional-Order Momentum
    Kan T.
    Gao Z.
    Yang C.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2020, 33 (06): : 559 - 567
  • [9] Accelerating gradient descent and Adam via fractional gradients
    Shin, Yeonjong
    Darbon, Jerome
    Karniadakis, George Em
    NEURAL NETWORKS, 2023, 161 : 185 - 201
  • [10] Improving pattern recognition accuracy of partial discharges by new data preprocessing methods
    Majidi, Mehrdad
    Oskuoee, Mohammad
    ELECTRIC POWER SYSTEMS RESEARCH, 2015, 119 : 100 - 110