Improving the Accuracy of Neural Network Pattern Recognition by Fractional Gradient Descent

被引:1
|
作者
Abdulkadirov, Ruslan I. [1 ]
Lyakhov, Pavel A. [1 ]
Baboshina, Valentina A. [1 ]
Nagornov, Nikolay N. [2 ]
机构
[1] North Caucasus Fed Univ, North Caucasus Ctr Math Res, Stavropol 355017, Russia
[2] North Caucasus Fed Univ, Dept Math Modeling, Stavropol 355017, Russia
来源
IEEE ACCESS | 2024年 / 12卷
基金
俄罗斯科学基金会;
关键词
Neural networks; Optimization; Pattern recognition; Accuracy; Convergence; Training; Transformers; Heuristic algorithms; Stability analysis; Multilayer perceptrons; Convolutional neural networks; fractional derivatives of Riemann-Liouville; Caputo; Grunwald-Letnikov; multilayer perceptron; optimization algorithms; stochastic gradient descent; OPTIMIZATION;
D O I
10.1109/ACCESS.2024.3491614
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper we propose the fractional gradient descent for increasing the training and work of modern neural networks. This optimizer searches the global minimum of the loss function considering the fractional gradient directions achieved by Riemann-Liouville, Caputo, and Grunwald-Letnikov derivatives. The adjusting of size and direction of the fractional gradient, supported by momentum and Nesterov condition, let the proposed optimizer descend into the global minimum of loss functions of neural networks. Utilizing the proposed optimization algorithm in a linear neural network and a visual transformer lets us attain higher accuracy, precision, recall, Macro F1 score by 1.8-4 percentage points than known analogs than state-of-the-art methods in solving pattern recognition problems on images from MNIST and CIFAR10 datasets. Further research of fractional calculus in modern neural network methodology can improve the quality of solving various challenges such as pattern recognition, time series forecasting, moving object detection, and data generation.
引用
收藏
页码:168428 / 168444
页数:17
相关论文
共 50 条
  • [31] Convergence Analysis and Application for Multi-Layer Neural Network Based on Fractional-Order Gradient Descent Learning
    Zhao, Shuai
    Fan, Qinwei
    Dong, Qingmei
    ADVANCED THEORY AND SIMULATIONS, 2024, 7 (02)
  • [32] A neural network applied to pattern recognition in statistical process control
    Anagun, AS
    COMPUTERS & INDUSTRIAL ENGINEERING, 1998, 35 (1-2) : 185 - 188
  • [33] A Novel Event-Driven Spiking Convolutional Neural Network for Electromyography Pattern Recognition
    Xu, Mengjuan
    Chen, Xiang
    Sun, Antong
    Zhang, Xu
    Chen, Xun
    IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2023, 70 (09) : 2604 - 2615
  • [34] Interpreting Neural Network Pattern With Pruning for PolSAR Target Recognition
    Lin, Huiping
    Yin, Junjun
    Yang, Jian
    Xu, Feng
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62
  • [35] Initialization-Based k-Winners-Take-All Neural Network Model Using Modified Gradient Descent
    Zhang, Yinyan
    Li, Shuai
    Geng, Guanggang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (08) : 4130 - 4138
  • [36] Stochastic gradient descent analysis for the evaluation of a speaker recognition
    Ashrf Nasef
    Marina Marjanović-Jakovljević
    Angelina Njeguš
    Analog Integrated Circuits and Signal Processing, 2017, 90 : 389 - 397
  • [37] Pattern Recognition Based on Heuristic Neural Network
    Su, Kuo-Ho
    Lee, Tan-Feng
    Liao, Hao Yu
    2017 INTERNATIONAL CONFERENCE ON SYSTEM SCIENCE AND ENGINEERING (ICSSE), 2017, : 284 - 288
  • [38] Multimodal pattern recognition by modular neural network
    Yang, SL
    Chang, KC
    OPTICAL ENGINEERING, 1998, 37 (02) : 650 - 659
  • [39] Improving the predictive accuracy of artificial neural network (ANN) approach in a mild steel turning operation
    Sada, Samuel O.
    INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2021, 112 (9-10) : 2389 - 2398