Convergence analysis of gradient descent stochastic algorithms

被引:54
|
作者
Shapiro, A [1 ]
Wardi, Y [1 ]
机构
[1] GEORGIA INST TECHNOL,SCH ELECT & COMP ENGN,ATLANTA,GA 30332
关键词
gradient descent; subdifferentials; uniform laws of large numbers; infinitesimal perturbation analysis; discrete event dynamic systems;
D O I
10.1007/BF02190104
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
This paper proves convergence of a sample-path based stochastic gradient-descent algorithm for optimizing expected-value performance measures in discrete event systems. The algorithm uses increasing precision at successive iterations, and it moves against the direction of a generalized gradient of the computed sample performance function. Two convergence results are established: one, for the case where the expected-value function is continuously differentiable; and the other, when that function is nondifferentiable but the sample performance functions are convex. The proofs are based on a version of the uniform law of large numbers which is provable for many discrete event systems where infinitesimal perturbation analysis is known to be strongly consistent.
引用
收藏
页码:439 / 454
页数:16
相关论文
共 50 条
  • [31] Convergence Analysis of Distributed Gradient Descent Algorithms With One and Two Momentum Terms
    Liu, Bing
    Chai, Li
    Yi, Jingwen
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (03) : 1511 - 1522
  • [32] New convergence aspects of stochastic gradient algorithms
    Nguyen, Lam M.
    Nguyen, Phuong Ha
    Richtárik, Peter
    Scheinberg, Katya
    Takáč, Martin
    van Dijk, Marten
    Journal of Machine Learning Research, 2019, 20
  • [33] New Convergence Aspects of Stochastic Gradient Algorithms
    Nguyen, Lam M.
    Phuong Ha Nguyen
    Richtarik, Peter
    Scheinberg, Katya
    Takac, Martin
    van Dijk, Marten
    JOURNAL OF MACHINE LEARNING RESEARCH, 2019, 20
  • [34] Convergence diagnostics for stochastic gradient descent with constant learning rate
    Chee, Jerry
    Toulis, Panos
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [35] A simplified convergence theory for Byzantine resilient stochastic gradient descent
    Roberts, Lindon
    Smyth, Edward
    EURO JOURNAL ON COMPUTATIONAL OPTIMIZATION, 2022, 10
  • [36] Fast Convergence for Stochastic and Distributed Gradient Descent in the Interpolation Limit
    Mitra, Partha P.
    2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 1890 - 1894
  • [37] Comparing Stochastic Gradient Descent and Mini-batch Gradient Descent Algorithms in Loan Risk Assessment
    Adigun, Abodunrin AbdulGafar
    Yinka-Banjo, Chika
    INFORMATICS AND INTELLIGENT APPLICATIONS, 2022, 1547 : 283 - 296
  • [38] Convergence Analysis of Gradient Descent for Eigenvector Computation
    Xu, Zhiqiang
    Cao, Xin
    Gao, Xin
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 2933 - 2939
  • [39] Simple Stochastic and Online Gradient Descent Algorithms for Pairwise Learning
    Yang, Zhenhuan
    Lei, Yunwen
    Wang, Puyu
    Yang, Tianbao
    Ying, Yiming
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [40] Convergence analysis of stochastic algorithms
    Shapiro, A
    Wardi, Y
    MATHEMATICS OF OPERATIONS RESEARCH, 1996, 21 (03) : 615 - 628