Central limit theorems for stochastic gradient descent with averaging for stable manifolds*

被引:2
|
作者
Dereich, Steffen [1 ]
Kassing, Sebastian [2 ]
机构
[1] Univ Munster, Inst Math Stochast, Fac Math & Comp Sci, Munster, Germany
[2] Univ Bielefeld, Fac Math, Bielefeld, Germany
来源
ELECTRONIC JOURNAL OF PROBABILITY | 2023年 / 28卷
关键词
stochastic approximation; Robbins-Monro; Ruppert-Polyak average; deep learning; stable manifold; APPROXIMATION;
D O I
10.1214/23-EJP947
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In this article, we establish new central limit theorems for Ruppert-Polyak averaged stochastic gradient descent schemes. Compared to previous work we do not assume that convergence occurs to an isolated attractor but instead allow convergence to a stable manifold. On the stable manifold the target function is constant and the oscillations of the iterates in the tangential direction may be significantly larger than the ones in the normal direction. We still recover a central limit theorem for the averaged scheme in the normal direction with the same rates as in the case of isolated attractors. In the setting where the magnitude of the random perturbation is of constant order, our research covers step-sizes -yn = C gamma n-gamma with C gamma > 0 and -y is an element of (34, 1). In particular, we show that the beneficial effect of averaging prevails in more general situations.
引用
收藏
页数:48
相关论文
共 50 条