Looking at the posterior: accuracy and uncertainty of neural-network predictions

被引:1
作者
Linander, Hampus [1 ,2 ]
Balabanov, Oleksandr [3 ]
Yang, Henry [1 ]
Mehlig, Bernhard [1 ]
机构
[1] Univ Gothenburg, Dept Phys, S-41296 Gothenburg, Sweden
[2] Univ Gothenburg, Chalmers Univ Technol, Dept Math Sci, S-41296 Gothenburg, Sweden
[3] Stockholm Univ, Dept Phys, S-10691 Stockholm, Sweden
来源
MACHINE LEARNING-SCIENCE AND TECHNOLOGY | 2023年 / 4卷 / 04期
基金
瑞典研究理事会;
关键词
deep learning; uncertainty quantification; bayesian inference; neural networks; active learning; QUANTIFICATION;
D O I
10.1088/2632-2153/ad0ab4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Bayesian inference can quantify uncertainty in the predictions of neural networks using posterior distributions for model parameters and network output. By looking at these posterior distributions, one can separate the origin of uncertainty into aleatoric and epistemic contributions. One goal of uncertainty quantification is to inform on prediction accuracy. Here we show that prediction accuracy depends on both epistemic and aleatoric uncertainty in an intricate fashion that cannot be understood in terms of marginalized uncertainty distributions alone. How the accuracy relates to epistemic and aleatoric uncertainties depends not only on the model architecture, but also on the properties of the dataset. We discuss the significance of these results for active learning and introduce a novel acquisition function that outperforms common uncertainty-based methods. To arrive at our results, we approximated the posteriors using deep ensembles, for fully-connected, convolutional and attention-based neural networks.
引用
收藏
页数:16
相关论文
共 54 条
[1]   A review of uncertainty quantification in deep learning: Techniques, applications and challenges [J].
Abdar, Moloud ;
Pourpanah, Farhad ;
Hussain, Sadiq ;
Rezazadegan, Dana ;
Liu, Li ;
Ghavamzadeh, Mohammad ;
Fieguth, Paul ;
Cao, Xiaochun ;
Khosravi, Abbas ;
Acharya, U. Rajendra ;
Makarenkov, Vladimir ;
Nahavandi, Saeid .
INFORMATION FUSION, 2021, 76 :243-297
[2]  
Angluin D., 1988, Machine Learning, V2, P319, DOI 10.1007/BF00116828
[3]  
[Anonymous], 2012, Active Learning
[4]  
Atlas Les, 1989, Advances in neural information processing systems, V2
[5]   Denoising images corrupted with impulse, Gaussian, or a mixture of impulse and Gaussian noise [J].
Awad, Ali .
ENGINEERING SCIENCE AND TECHNOLOGY-AN INTERNATIONAL JOURNAL-JESTECH, 2019, 22 (03) :746-753
[6]   Bayesian posterior approximation with stochastic ensembles [J].
Balabanov, Oleksandr ;
Mehlig, Bernhard ;
Linander, Hampus .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, :13701-13711
[7]   The power of ensembles for active learning in image classification [J].
Beluch, William H. ;
Genewein, Tim ;
Nuernberger, Andreas ;
Koehler, Jan M. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9368-9377
[8]   Interpretable, calibrated neural networks for analysis and understanding of inelastic neutron scattering data [J].
Butler, Keith T. ;
Le, Manh Duc ;
Thiyagalingam, Jeyan ;
Perring, Toby G. .
JOURNAL OF PHYSICS-CONDENSED MATTER, 2021, 33 (19)
[9]  
Chitta K, 2019, Arxiv, DOI arXiv:1811.03575
[10]  
Dagli R., 2021, Image classification with swin-transformers