On Almost Sure Convergence Rates of Stochastic Gradient Methods

被引:0
作者
Liu, Jun [1 ]
Yuan, Ye [2 ,3 ]
机构
[1] Univ Waterloo, Dept Appl Math, Waterloo, ON, Canada
[2] Huazhong Univ Sci & Technol, Sch Artificial Intelligence & Automat, Wuhan, Peoples R China
[3] Huazhong Univ Sci & Technol, Sch Mech Sci & Engn, Wuhan, Peoples R China
来源
CONFERENCE ON LEARNING THEORY, VOL 178 | 2022年 / 178卷
基金
加拿大自然科学与工程研究理事会;
关键词
Stochastic gradient descent; stochastic heavy-ball; stochastic Nesterov's accelerated gradient; almost sure convergence rate; OPTIMIZATION; BOUNDS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The vast majority of convergence rates analysis for stochastic gradient methods in the literature focus on convergence in expectation, whereas trajectory-wise almost sure convergence is clearly important to ensure that any instantiation of the stochastic algorithms would converge with probability one. Here we provide a unified almost sure convergence rates analysis for stochastic gradient descent (SGD), stochastic heavy-ball (SHB), and stochastic Nesterov's accelerated gradient (SNAG) methods. We show, for the first time, that the almost sure convergence rates obtained for these stochastic gradient methods on strongly convex functions, are arbitrarily close to their optimal convergence rates possible. For non-convex objective functions, we not only show that a weighted average of the squared gradient norms converges to zero almost surely, but also the last iterates of the algorithms. We further provide last-iterate almost sure convergence rates analysis for stochastic gradient methods on general convex smooth functions, in contrast with most existing results in the literature that only provide convergence in expectation for a weighted average of the iterates.
引用
收藏
页数:21
相关论文
共 50 条
  • [31] Convergence behavior of diffusion stochastic gradient descent algorithm
    Barani, Fatemeh
    Savadi, Abdorreza
    Yazdi, Hadi Sadoghi
    SIGNAL PROCESSING, 2021, 183
  • [32] On the Convergence of Decentralized Stochastic Gradient Descent With Biased Gradients
    Jiang, Yiming
    Kang, Helei
    Liu, Jinlan
    Xu, Dongpo
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2025, 73 : 549 - 558
  • [33] Stochastic Gradient Methods with Preconditioned Updates
    Abdurakhmon Sadiev
    Aleksandr Beznosikov
    Abdulla Jasem Almansoori
    Dmitry Kamzolov
    Rachael Tappenden
    Martin Takáč
    Journal of Optimization Theory and Applications, 2024, 201 : 471 - 489
  • [34] Stochastic Gradient Methods with Preconditioned Updates
    Sadiev, Abdurakhmon
    Beznosikov, Aleksandr
    Almansoori, Abdulla Jasem
    Kamzolov, Dmitry
    Tappenden, Rachael
    Takac, Martin
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2024, 201 (02) : 471 - 489
  • [35] Novel Convergence Results of Adaptive Stochastic Gradient Descents
    Sun, Tao
    Qiao, Linbo
    Liao, Qing
    Li, Dongsheng
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 1044 - 1056
  • [36] Convergence rates analysis of a multiobjective proximal gradient method
    Tanabe, Hiroki
    Fukuda, Ellen H.
    Yamashita, Nobuo
    OPTIMIZATION LETTERS, 2023, 17 (02) : 333 - 350
  • [37] Almost Sure Exponential Stability of Uncertain Stochastic Hopfield Neural Networks Based on Subadditive Measures
    Jia, Zhifu
    Li, Cunlin
    MATHEMATICS, 2023, 11 (14)
  • [38] On the Convergence of Nested Decentralized Gradient Methods With Multiple Consensus and Gradient Steps
    Berahas, Albert S.
    Bollapragada, Raghu
    Wei, Ermin
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 (69) : 4192 - 4203
  • [39] On the Convergence of Nesterov's Accelerated Gradient Method in Stochastic Settings
    Assran, Mahmoud
    Rabbat, Michael
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [40] CEDAS: A Compressed Decentralized Stochastic Gradient Method With Improved Convergence
    Huang, Kun
    Pu, Shi
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2025, 70 (04) : 2242 - 2257