Improving predictive uncertainty estimation using Dropout-Hamiltonian Monte Carlo

被引:9
作者
Hernandez, Sergio [1 ]
Vergara, Diego [1 ]
Valdenegro-Toro, Matias [2 ]
Jorquera, Felipe [1 ]
机构
[1] Univ Catolica Maule, Ctr Innovac Ingn Aplicada, Talca, Chile
[2] German Res Ctr Artificial Intelligence, Robot Innovat Ctr, Bremen, Germany
关键词
Bayesian learning; Hamiltonian Monte Carlo; Dropout; Transfer learning; Classification; LANGEVIN;
D O I
10.1007/s00500-019-04195-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Estimating predictive uncertainty is crucial for many computer vision tasks, from image classification to autonomous driving systems. Hamiltonian Monte Carlo (HMC) is an sampling method for performing Bayesian inference. On the other hand, Dropout regularization has been proposed as an approximate model averaging technique that tends to improve generalization in large-scale models such as deep neural networks. Although HMC provides convergence guarantees for most standard Bayesian models, it do not handle discrete parameters arising from Dropout regularization. In this paper, we present a robust methodology for improving predictive uncertainty in classification problems, based on Dropout and HMC. Even though Dropout induces a non-smooth energy function with no such convergence guarantees, the resulting discretization of the Hamiltonian proves empirical success. The proposed method allows to effectively estimate the predictive accuracy and to provide better generalization for difficult test examples.
引用
收藏
页码:4307 / 4322
页数:16
相关论文
共 39 条
[1]  
Afshar HM, 2015, ADV NEUR IN, V28
[2]  
[Anonymous], 2013, ADV NEURAL INFORM PR
[3]  
[Anonymous], 2013, ARXIV13126197
[4]  
[Anonymous], 2017, C EMPIRICAL METHODSI
[5]  
[Anonymous], 2017, ARXIV170508510
[6]  
[Anonymous], 2013, Bayesian Data Analysis
[7]  
[Anonymous], 2012, Computer vision: models, learning, and inference
[8]  
[Anonymous], Evaluating Derivatives
[9]  
[Anonymous], 2013, INT C MACH LEARN
[10]  
[Anonymous], 2011, HDB MARKOV CHAIN MON