Overfitting Measurement of Deep Neural Networks Using No Data

被引:8
作者
Watanabe, Satoru [1 ]
Yamana, Hayato [2 ]
机构
[1] Waseda Univ, Dept Comp Sci & Commun Engn, Tokyo, Japan
[2] Waseda Univ, Fac Sci & Engn, Tokyo, Japan
来源
2021 IEEE 8TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA) | 2021年
关键词
deep neural network; overfitting; persistent homology;
D O I
10.1109/DSAA53316.2021.9564119
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Overfitting reduces the generalizability of deep neural networks (DNNs). Overfitting is generally detected by comparing the accuracies and losses of training and validation data; however, the detection method requires vast amounts of training data and is not always effective for forthcoming data due to the heterogeneity between training and forthcoming data. The dropout technique has been employed to prevent DNNs from overfitting, where the neurons in DNNs are invalidated randomly during their training. It has been hypothesized that this technique prevents DNNs from overfitting by restraining the co-adaptions among neurons. This hypothesis implies that overfitting of a DNN is a result of the co-adaptions among neurons and can be detected by investigating the inner representation of DNNs. Thus, we propose a method to detect overfitting of DNNs using no training and test data. The proposed method measures the degree of co-adaptions among neurons using persistent homology (PH). The proposed PH-based overfitting measure (PHOM) method constructs clique complexes on DNNs using the trained parameters of DNNs, and the one-dimensional PH investigates the co-adaptions among neurons. Thus, PHOM requires no training and test data to measure overfitting. We applied PHOM to convolutional neural networks trained for the classification problems of the CIFAR-10, SVHN, and Tiny ImageNet data sets. The experimental results demonstrate that PHOM reveals the degree of overfitting of DNNs to the training data, which suggests that PHOM enables us to filter overfitted DNNs without requiring the training and test data.
引用
收藏
页数:10
相关论文
共 34 条
  • [1] Adams Henry, 2014, Mathematical Software - ICMS 2014. 4th International Congress. Proceedings. LNCS: 8592, P129, DOI 10.1007/978-3-662-44199-2_23
  • [2] Bishop Christopher M., 2006, Pattern recognition and machine learning, V1st edn
  • [3] On the Reliability of Individual Brain Activity Networks
    Cassidy, Ben
    Bowman, F. DuBois
    Rae, Caroline
    Solo, Victor
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2018, 37 (02) : 649 - 662
  • [4] Chollet F., 2018, DEEP LEARNING PYTHON, V361
  • [5] Computing the Testing Error without a Testing Set
    Corneanu, Ciprian A.
    Escalera, Sergio
    Martinez, Aleix M.
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 2674 - 2682
  • [6] What does it mean to learn in deep networks? And, how does one detect adversarial attacks?
    Corneanu, Ciprian A.
    Madadi, Meysam
    Escalera, Sergio
    Martinez, Aleix M.
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4752 - 4761
  • [7] Dlotko P., 2015, GUDHI user and reference manual
  • [8] Edelsbrunner H, 2000, ANN IEEE SYMP FOUND, P454
  • [9] Edelsbrunner H., 2012, Technical report
  • [10] Edelsbrunner H., 2010, COMPUTATIONAL TOPOLO, DOI DOI 10.1007/978-3-540-33259-6_7