Convergence analysis of gradient descent stochastic algorithms

被引:54
|
作者
Shapiro, A [1 ]
Wardi, Y [1 ]
机构
[1] GEORGIA INST TECHNOL,SCH ELECT & COMP ENGN,ATLANTA,GA 30332
关键词
gradient descent; subdifferentials; uniform laws of large numbers; infinitesimal perturbation analysis; discrete event dynamic systems;
D O I
10.1007/BF02190104
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
This paper proves convergence of a sample-path based stochastic gradient-descent algorithm for optimizing expected-value performance measures in discrete event systems. The algorithm uses increasing precision at successive iterations, and it moves against the direction of a generalized gradient of the computed sample performance function. Two convergence results are established: one, for the case where the expected-value function is continuously differentiable; and the other, when that function is nondifferentiable but the sample performance functions are convex. The proofs are based on a version of the uniform law of large numbers which is provable for many discrete event systems where infinitesimal perturbation analysis is known to be strongly consistent.
引用
收藏
页码:439 / 454
页数:16
相关论文
共 50 条