Nonasymptotic analysis of Stochastic Gradient Hamiltonian Monte Carlo under local conditions for nonconvex optimization

被引:0
作者
Akyildiz, O. Deniz [1 ]
Sabanis, Sotirios [2 ,3 ]
机构
[1] Imperial Coll London, Dept Math, London, England
[2] Univ Edinburgh, Sch Math, Edinburgh, Scotland
[3] Natl Tech Univ Athens, Alan Turing Inst, Athens, Greece
基金
英国工程与自然科学研究理事会;
关键词
Non-convex optimization; underdamped Langevin Monte Carlo; non-log-concave sampling; nonasmyptotic bounds; global optimization; DEPENDENT DATA STREAMS; LANGEVIN DYNAMICS; CONVERGENCE;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We provide a nonasymptotic analysis of the convergence of the stochastic gradient Hamiltonian Monte Carlo (SGHMC) to a target measure in Wasserstein-2 distance without assuming log-concavity. Our analysis quantifies key theoretical properties of the SGHMC as a sampler under local conditions which significantly improves the findings of previous results. In particular, we prove that the Wasserstein-2 distance between the target and the law of the SGHMC is uniformly controlled by the step-size of the algorithm, therefore demonstrate that the SGHMC can provide high-precision results uniformly in the number of iterations. The analysis also allows us to obtain nonasymptotic bounds for nonconvex optimization problems under local conditions and implies that the SGHMC, when viewed as a nonconvex optimizer, converges to a global minimum with the best known rates. We apply our results to obtain nonasymptotic bounds for scalable Bayesian inference and nonasymptotic generalization bounds.
引用
收藏
页数:34
相关论文
共 42 条
[1]  
Ahn Sungjin, 2012, P 29 INT COFERENCE I, DOI DOI 10.5555/3042573.3042799
[2]  
Balasubramanian K, 2022, PR MACH LEARN RES, V178
[3]   On stochastic gradient Langevin dynamics with dependent data streams in the logconcave case [J].
Barkhagen, M. ;
Chau, N. H. ;
Moulines, E. ;
Rasonyi, M. ;
Sabanis, S. ;
Zhang, Y. .
BERNOULLI, 2021, 27 (01) :1-33
[4]  
Brosse N., 2018, Advances in Neural Information Processing Systems, V31, P8268
[5]   The tamed unadjusted Langevin algorithm [J].
Brosse, Nicolas ;
Durmus, Alain ;
Moulines, Eric ;
Sabanis, Sotirios .
STOCHASTIC PROCESSES AND THEIR APPLICATIONS, 2019, 129 (10) :3638-3663
[6]   Stochastic Gradient Hamiltonian Monte Carlo for non-convex learning [J].
Chau, Huy N. ;
Rasonyi, Miklos .
STOCHASTIC PROCESSES AND THEIR APPLICATIONS, 2022, 149 :341-368
[7]   ON FIXED GAIN RECURSIVE ESTIMATORS WITH DISCONTINUITY IN THE PARAMETERS [J].
Chau, Huy N. ;
Kumar, Chaman ;
Rasonyi, Miklos ;
Sabanis, Sotirios .
ESAIM-PROBABILITY AND STATISTICS, 2019, 23 :217-244
[8]   On Stochastic Gradient Langevin Dynamics with Dependent Data Streams: The Fully Nonconvex Case [J].
Chau, Ngoc Huy ;
Moulines, Eric ;
Rasonyi, Miklos ;
Sabanis, Sotirios ;
Zhang, Ying .
SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2021, 3 (03) :959-986
[9]  
Chen TQ, 2014, PR MACH LEARN RES, V32, P1683
[10]  
Cheng X, 2020, Arxiv, DOI arXiv:1805.01648