An adaptively weighted stochastic gradient MCMC algorithm for Monte Carlo simulation and global optimization

被引:7
作者
Deng, Wei [1 ]
Lin, Guang [2 ]
Liang, Faming [2 ,3 ]
机构
[1] Purdue Univ, Dept Stat, W Lafayette, IN 47907 USA
[2] Purdue Univ, Dept Math, W Lafayette, IN 47907 USA
[3] Purdue Univ, Sch Mech Engn, W Lafayette, IN 47907 USA
基金
美国国家科学基金会; 美国国家卫生研究院;
关键词
Adaptive stochastic gradient Langevin dynamics; Dynamic importance sampling; Local traps; Stochastic approximation; APPROXIMATION; CONVERGENCE; LANGEVIN;
D O I
10.1007/s11222-022-10120-3
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
We propose an adaptively weighted stochastic gradient Langevin dynamics (AWSGLD) algorithm for Bayesian learning of big data problems. The proposed algorithm is scalable and possesses a self-adjusting mechanism: It adaptively flattens the high-energy region and protrudes the low-energy region during simulations such that both Monte Carlo simulation and global optimization tasks can be greatly facilitated in a single run. The self-adjusting mechanism enables the proposed algorithm to be essentially immune to local traps. Theoretically, by showing the stability of the mean-field system and verifying the existence and regularity properties of the solution of Poisson equation, we establish the convergence of the AWSGLD algorithm, including both the convergence of the self-adapting parameters and the convergence of the weighted averaging estimators. Empirically, the AWSGLD algorithm is tested on multiple benchmark datasets including CIFAR100 and SVHN for both optimization and uncertainty estimation tasks. The numerical results indicate its great potential in Monte Carlo simulation and global optimization for modern machine learning tasks.
引用
收藏
页数:24
相关论文
共 67 条
[11]  
Chen TQ, 2014, PR MACH LEARN RES, V32, P1683
[12]  
Chen Y., 2019, INT C LEARNING REPRE
[13]  
Deng W., 2021, INT C LEARNING REPRE
[14]  
Deng W., 2020, Advances in Neural Information Processing Systems (NeurIPS)
[15]  
Deng Wei, 2020, Proc Mach Learn Res, V119, P2474
[16]  
Ding N., 2014, Advances in Neural Information Processing Systems, V27
[17]  
Erdogdu M.A., 2018, Advances in Neural Information Processing Systems (NeurIPS)
[18]   CONVERGENCE OF THE WANG-LANDAU ALGORITHM [J].
Fort, Gersende ;
Jourdain, Benjamin ;
Kuhn, Estelle ;
Lelievre, Tony ;
Stoltz, Gabriel .
MATHEMATICS OF COMPUTATION, 2015, 84 (295) :2297-2327
[19]   ANNEALING MARKOV-CHAIN MONTE-CARLO WITH APPLICATIONS TO ANCESTRAL INFERENCE [J].
GEYER, CJ ;
THOMPSON, EA .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1995, 90 (431) :909-920
[20]  
GEYER CJ, 1991, COMPUTING SCIENCE AND STATISTICS, P156