The Effect of Training Data Quantity on Monte Carlo Dropout Uncertainty Quantification in Deep Learning

被引:3
作者
Cusack, Harrison [1 ]
Bialkowski, Alina [1 ]
机构
[1] Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld, Australia
来源
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN | 2023年
关键词
D O I
10.1109/IJCNN54540.2023.10191327
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
When deploying deep neural networks, quantification of a model's uncertainty is necessary to provide confidence in its predictions by distinguishing between accurate predictions and coincidentally correct guesses. While it is known that the accuracy of predictions is dependent on the data on which the model was trained, to date, limited work has examined the relationship between training data quantity and uncertainty quantification. In this paper, we propose two metrics to assess the 'quality' of uncertainty quantification, and investigate the relationship between training data quantity and Monte Carlo Dropout uncertainty quantification in supervised and semisupervised learning across various text-based datasets. We found that in supervised learning, uncertainty quantification quality (across both metrics) initially increased for larger quantities of training data, but interestingly, after a certain threshold, began to gradually decline. In semi-supervised learning, uncertainty quantification was enhanced by both a greater number of training samples and greater proportion of pre-labelled data. These results suggest that for supervised learning, data scientists generally ought not to invest resources into acquiring more training data solely for superior uncertainty quantification. However, if semi-supervised learning is necessary, then there is a marked benefit in obtaining more data.
引用
收藏
页数:8
相关论文
共 33 条
  • [1] A review of uncertainty quantification in deep learning: Techniques, applications and challenges
    Abdar, Moloud
    Pourpanah, Farhad
    Hussain, Sadiq
    Rezazadegan, Dana
    Liu, Li
    Ghavamzadeh, Mohammad
    Fieguth, Paul
    Cao, Xiaochun
    Khosravi, Abbas
    Acharya, U. Rajendra
    Makarenkov, Vladimir
    Nahavandi, Saeid
    [J]. INFORMATION FUSION, 2021, 76 : 243 - 297
  • [2] Illustrative Discussion of MC-Dropout in General Dataset: Uncertainty Estimation in Bitcoin
    Alarab, Ismail
    Prakoonwit, Simant
    Nacer, Mohamed Ikbal
    [J]. NEURAL PROCESSING LETTERS, 2021, 53 (02) : 1001 - 1011
  • [3] [Anonymous], 2002, Proceedings of Flairs-02
  • [4] Calderon-Ramirez Saul., 2021, 2021 Int. Joint Conf. on Neural Networks (IJCNN), P1
  • [5] Chapelle O., 2009, Semi-supervised Learning, V20, P542
  • [6] Dai Andrew M., 2015, Semi-supervised sequence learning. Ad
  • [7] Davidson T, 2017, P INT AAAI C WEB SOC, P1, DOI [DOI 10.1609/ICWSM.V11I1.14955, 10.1609/icwsm.v11i1.14955]
  • [8] Data Labeling: An Empirical Investigation into Industrial Challenges and Mitigation Strategies
    Fredriksson, Teodor
    Mattos, David Issa
    Bosch, Jan
    Olsson, Helena Holmstrom
    [J]. PRODUCT-FOCUSED SOFTWARE PROCESS IMPROVEMENT (PROFES 2020), 2020, 12562 : 202 - 216
  • [9] Gal Y., 2017, ADV NEURAL INFORM PR, V30, P10
  • [10] Gal Y, 2016, PR MACH LEARN RES, V48