On the least amount of training data for a machine learning model

被引:1
作者
Zhao, Dazhi [1 ,2 ]
Hao, Yunquan [1 ]
Li, Weibin [3 ]
Tu, Zhe [4 ]
机构
[1] Southwest Petr Univ, Sch Sci, Chengdu, Peoples R China
[2] Southwest Petr Univ, Inst Artificial Intelligence, Chengdu, Peoples R China
[3] China Aerodynam Res & Dev Ctr, Mianyang 621000, Sichuan, Peoples R China
[4] Zhejiang Wanli Univ, Coll Big Data & Software Engn, Ningbo, Peoples R China
基金
浙江省自然科学基金;
关键词
Machine learning; sampling theorem; frequency principle; signal recovery; neural network; Gaussian process regression; DEEP NEURAL-NETWORKS;
D O I
10.3233/JIFS-211024
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Whether the exact amount of training data is enough for a specific task is an important question in machine learning, since it is always very expensive to label many data while insufficient data lead to underfitting. In this paper, the topic that what is the least amount of training data for a model is discussed from the perspective of sampling theorem. If the target function of supervised learning is taken as a multi-dimensional signal and the labeled data as samples, the training process can be regarded as the process of signal recovery. The main result is that the least amount of training data for a bandlimited task signal corresponds to a sampling rate which is larger than the Nyquist rate. Some numerical experiments are carried out to show the comparison between the learning process and the signal recovery, which demonstrates our result. Based on the equivalence between supervised learning and signal recovery, some spectral methods can be used to reveal underlying mechanisms of various supervised learning models, especially those "black-box" neural networks.
引用
收藏
页码:4891 / 4906
页数:16
相关论文
共 34 条
  • [1] [Anonymous], 1983, Signals and systems
  • [2] Cai W, 2019, Arxiv, DOI arXiv:1905.01389
  • [3] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [4] A fourier-based approach to generalization and optimization in deep learning
    Farnia F.
    Zhang J.M.
    Tse D.N.
    [J]. IEEE Journal on Selected Areas in Information Theory, 2020, 1 (01): : 145 - 156
  • [5] Incremental Missing-Data Imputation for Evolving Fuzzy Granular Prediction
    Garcia, Cristiano
    Leite, Daniel
    Skrjanc, Igor
    [J]. IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2020, 28 (10) : 2348 - 2362
  • [6] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [7] Hinton G.E., 2012, arXiv, DOI DOI 10.48550/ARXIV.1207.0580
  • [8] Intelligent Detection of the PV Faults Based on Artificial Neural Network and Type 2 Fuzzy Systems
    Janarthanan, Ramadoss
    Maheshwari, R. Uma
    Shukla, Prashant Kumar
    Shukla, Piyush Kumar
    Mirjalili, Seyedali
    Kumar, Manoj
    [J]. ENERGIES, 2021, 14 (20)
  • [9] Regularization of deep neural networks with spectral dropout
    Khan, Salman H.
    Hayat, Munawar
    Porikli, Fatih
    [J]. NEURAL NETWORKS, 2019, 110 : 82 - 90
  • [10] Krizhevsky A., 2012, Advances in Neural Information Processing Systems, P1097, DOI DOI 10.1145/3065386