A stochastic recursive gradient algorithm integrating momentum and the powerball function with adaptive step sizes

被引:0
|
作者
Qin, Chuandong [1 ,2 ]
Cai, Zilin [1 ]
Guo, Yuhang [1 ]
机构
[1] North Minzu Univ, Sch Math & Informat Sci, 204 Wenchang North St, Yinchuan 750021, Ningxia, Peoples R China
[2] North Minzu Univ, Ningxia Key Lab Intelligent Informat & Big Data Pr, 204 Wenchang North St, Yinchuan 750021, Ningxia, Peoples R China
基金
中国国家自然科学基金;
关键词
Machine learning; Variance reduction; Momentum; Powerball function; Adaptive learning rate; DESCENT; MINIMIZATION;
D O I
10.1007/s13042-024-02514-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Momentum techniques and the Powerball function have been proven effective in stochastic optimization algorithms, widely utilized in large-scale optimization scenarios. Nonetheless, the integration of these methodologies into stochastic optimization algorithms and the determination of their initial learning rates persist as unresolved and pivotal issues. In this study, we integrate momentum techniques and the Powerball function into the SARAH (StochAstic Recursive grAdient algoritHm), culminating in the inception of a novel variance-reduced gradient descent algorithm named PM-SARAH. Moreover, a pair of adaptive step size variants are respectively integrated into the outer and inner loops of PM-SARAH, giving rise to PM-SARAH-AS and PM-SARAH-RAS. Ultimately, through comparative experimentation with state-of-the-art optimization algorithms on standard machine learning tasks and certain non-convex scenarios, the empirical results underscore the superior performance of the algorithms elucidated in this paper.
引用
收藏
页数:21
相关论文
共 14 条
  • [1] Adaptive Powerball Stochastic Conjugate Gradient for Large-Scale Learning
    Yang, Zhuang
    IEEE TRANSACTIONS ON BIG DATA, 2023, 9 (06) : 1598 - 1606
  • [2] SARAH-M: A fast stochastic recursive gradient descent algorithm via momentum
    Yang, Zhuang
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [3] A stochastic gradient tracking algorithm with adaptive momentum for distributed optimization
    Li, Yantao
    Hu, Hanqing
    Zhang, Keke
    Lu, Qingguo
    Deng, Shaojiang
    Li, Huaqing
    NEUROCOMPUTING, 2025, 637
  • [4] Proximal stochastic recursive momentum algorithm for nonsmooth nonconvex optimization problems
    Wang, Zhaoxin
    Wen, Bo
    OPTIMIZATION, 2024, 73 (02) : 481 - 495
  • [5] Biased stochastic conjugate gradient algorithm with adaptive step size for nonconvex problems
    Huang, Ruping
    Qin, Yan
    Liu, Kejun
    Yuan, Gonglin
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [6] An improved adaptive momentum gradient descent algorithm
    Jiang Z.
    Song J.
    Liu Y.
    Huazhong Keji Daxue Xuebao (Ziran Kexue Ban)/Journal of Huazhong University of Science and Technology (Natural Science Edition), 2023, 51 (05): : 137 - 143
  • [7] ADINE: An Adaptive Momentum Method for Stochastic Gradient Descent
    Srinivasan, Vishwak
    Sankar, Adepu Ravi
    Balasubramanian, Vineeth N.
    PROCEEDINGS OF THE ACM INDIA JOINT INTERNATIONAL CONFERENCE ON DATA SCIENCE AND MANAGEMENT OF DATA (CODS-COMAD'18), 2018, : 249 - 256
  • [8] A new inexact stochastic recursive gradient descent algorithm with Barzilai–Borwein step size in machine learning
    Yi-ming Yang
    Fu-sheng Wang
    Jin-xiang Li
    Yuan-yuan Qin
    Nonlinear Dynamics, 2023, 111 : 3575 - 3586
  • [9] Adaptive Polyak Step-Size for Momentum Accelerated Stochastic Gradient Descent With General Convergence Guarantee
    Zhang, Jiawei
    Jin, Cheng
    Gu, Yuantao
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2025, 73 : 462 - 476
  • [10] A new inexact stochastic recursive gradient descent algorithm with Barzilai-Borwein step size in machine learning
    Yang, Yi-ming
    Wang, Fu-sheng
    Li, Jin-xiang
    Qin, Yuan-yuan
    NONLINEAR DYNAMICS, 2023, 111 (04) : 3575 - 3586