HatchEnsemble: an efficient and practical uncertainty quantification method for deep neural networks

被引:0
作者
Yufeng Xia
Jun Zhang
Tingsong Jiang
Zhiqiang Gong
Wen Yao
Ling Feng
机构
[1] National University of Defense Technology,College of Aerospace Science and Engineering
[2] Tsinghua University,Department of Computer Science and Technology
[3] Chinese Academy of Military Science,National Innovation Institute of Defense Technology
来源
Complex & Intelligent Systems | 2021年 / 7卷
关键词
Ensemble learning; Deep neural networks; Non-Bayesian method; Uncertainty quantification;
D O I
暂无
中图分类号
学科分类号
摘要
Quantifying predictive uncertainty in deep neural networks is a challenging and yet unsolved problem. Existing quantification approaches can be categorized into two lines. Bayesian methods provide a complete uncertainty quantification theory but are often not scalable to large-scale models. Along another line, non-Bayesian methods have good scalability and can quantify uncertainty with high quality. The most remarkable idea in this line is Deep Ensemble, but it is limited in practice due to its expensive computational cost. Thus, we propose HatchEnsemble to improve the efficiency and practicality of Deep Ensemble. The main idea is to use function-preserving transformations, ensuring HatchNets to inherit the knowledge learned by a single model called SeedNet. This process is called hatching, and HatchNet can be obtained by continuously widening the SeedNet. Based on our method, two different hatches are proposed, respectively, for ensembling the same and different architecture networks. To ensure the diversity of models, we also add random noises to parameters during hatching. Experiments on both clean and corrupted datasets show that HatchEnsemble can give a competitive prediction performance and better-calibrated uncertainty quantification in a shorter time compared with baselines.
引用
收藏
页码:2855 / 2869
页数:14
相关论文
共 40 条
  • [1] LeCun Y(2015)Deep learning Nature 521 436-444
  • [2] Bengio Y(2017)Imagenet classification with deep convolutional neural networks Commun ACM 60 84-90
  • [3] Hinton G(2012)Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups IEEE Signal Process Mag 29 82-97
  • [4] Krizhevsky A(2015)Predicting the sequence specificities of dna-and rna-binding proteins by deep learning Nat Biotechnol 33 831-838
  • [5] Sutskever I(1992)A practical bayesian framework for backpropagation networks Neural Comput 4 448-472
  • [6] Hinton GE(2017)Simple and scalable predictive uncertainty estimation using deep ensembles Adv Neural Inf Process Syst 30 6402-6413
  • [7] Hinton G(1999)Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods Adv Large Margin Classif 10 61-74
  • [8] Deng L(2011)Practical variational inference for neural networks Adv Neural Inf Process Syst 24 2348-2356
  • [9] Yu D(2014)Dropout: a simple way to prevent neural networks from overfitting J Mach Learn Res 15 1929-1958
  • [10] Dahl GE(1998)Gradient-based learning applied to document recognition Proc IEEE 86 2278-2324