The Improved Stochastic Fractional Order Gradient Descent Algorithm

被引:2
|
作者
Yang, Yang [1 ]
Mo, Lipo [1 ,2 ]
Hu, Yusen [1 ]
Long, Fei [3 ]
机构
[1] Beijing Technol & Business Univ, Sch Math & Stat, Beijing 100048, Peoples R China
[2] Beijing Technol & Business Univ, Sch Future Technol, Beijing 100048, Peoples R China
[3] Guizhou Inst Technol, Sch Artificial Intelligence & Elect Engn, Special Key Lab Artificial Intelligence & Intellig, Guiyang 550003, Peoples R China
关键词
machine learning; fractional calculus; stochastic gradient descent; convex optimization; online optimization; NEURAL-NETWORKS;
D O I
10.3390/fractalfract7080631
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
This paper mainly proposes some improved stochastic gradient descent (SGD) algorithms with a fractional order gradient for the online optimization problem. For three scenarios, including standard learning rate, adaptive gradient learning rate, and momentum learning rate, three new SGD algorithms are designed combining a fractional order gradient and it is shown that the corresponding regret functions are convergent at a sub-linear rate. Then we discuss the impact of the fractional order on the convergence and monotonicity and prove that the better performance can be obtained by adjusting the order of the fractional gradient. Finally, several practical examples are given to verify the superiority and validity of the proposed algorithm.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Dendrite morphological neurons trained by stochastic gradient descent
    Zamora, Erik
    Sossa, Humberto
    NEUROCOMPUTING, 2017, 260 : 420 - 431
  • [32] Optimized convergence of stochastic gradient descent by weighted averaging
    Hagedorn, Melinda
    Jarre, Florian
    OPTIMIZATION METHODS & SOFTWARE, 2024, 39 (04) : 699 - 724
  • [33] Semantics-Preserving Parallelization of Stochastic Gradient Descent
    Maleki, Saeed
    Musuvathi, Madanlal
    Mytkowicz, Todd
    2018 32ND IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS), 2018, : 224 - 233
  • [34] Adaptive Stochastic Gradient Descent (SGD) for erratic datasets
    Dagal, Idriss
    Tanrioven, Kursat
    Nayir, Ahmet
    Akin, Burak
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2025, 166
  • [35] Embedding Simulated Annealing within Stochastic Gradient Descent
    Fischetti, Matteo
    Stringher, Matteo
    OPTIMIZATION AND LEARNING, OLA 2021, 2021, 1443 : 3 - 13
  • [36] Adjusted stochastic gradient descent for latent factor analysis
    Li, Qing
    Xiong, Diwen
    Shang, Mingsheng
    INFORMATION SCIENCES, 2022, 588 : 196 - 213
  • [37] On a Relationship between Integral Compensation and Stochastic Gradient Descent
    Fujimoto, Yusuke
    Maruta, Ichiro
    Sugie, Toshiharu
    2017 56TH ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS OF JAPAN (SICE), 2017, : 343 - 343
  • [38] Continuous-time stochastic gradient descent for optimizing over the stationary distribution of stochastic differential equations
    Wang, Ziheng
    Sirignano, Justin
    MATHEMATICAL FINANCE, 2024, 34 (02) : 348 - 424
  • [39] Recent Advances in Stochastic Gradient Descent in Deep Learning
    Tian, Yingjie
    Zhang, Yuqi
    Zhang, Haibin
    MATHEMATICS, 2023, 11 (03)
  • [40] On Projected Stochastic Gradient Descent Algorithm with Weighted Averaging for Least Squares Regression
    Cohen, Kobi
    Nedic, Angelia
    Srikant, R.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2017, 62 (11) : 5974 - 5981