Dropout Rademacher complexity of deep neural networks

被引:1
|
作者
Wei GAO [1 ,2 ]
Zhi-Hua ZHOU [1 ,2 ]
机构
[1] National Key Laboratory for Novel Software Technology, Nanjing University
[2] Collaborative Innovation Center of Novel Software Technology and Industrialization,Nanjing University
基金
中国国家自然科学基金;
关键词
artificial intelligence; machine learning; deep learning; dropout; Rademacher complexity;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Great successes of deep neural networks have been witnessed in various real applications. Many algorithmic and implementation techniques have been developed; however, theoretical understanding of many aspects of deep neural networks is far from clear. A particular interesting issue is the usefulness of dropout,which was motivated from the intuition of preventing complex co-adaptation of feature detectors. In this paper,we study the Rademacher complexity of different types of dropouts, and our theoretical results disclose that for shallow neural networks(with one or none hidden layer) dropout is able to reduce the Rademacher complexity in polynomial, whereas for deep neural networks it can amazingly lead to an exponential reduction.
引用
收藏
页码:173 / 184
页数:12
相关论文
共 50 条
  • [41] Plug and Play Deep Convolutional Neural Networks
    Neary, Patrick
    Allan, Vicki
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE (ICAART), VOL 2, 2019, : 388 - 395
  • [42] Deep Learning for Epidemiologists: An Introduction to Neural Networks
    Serghiou, Stylianos
    Rough, Kathryn
    AMERICAN JOURNAL OF EPIDEMIOLOGY, 2023, 192 (11) : 1904 - 1916
  • [43] Data Dropout: Optimizing Training Data for Convolutional Neural Networks
    Wang, Tianyang
    Huan, Jun
    Li, Bo
    2018 IEEE 30TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2018, : 39 - 46
  • [44] Deep Neural Networks Reveal a Gradient in the Complexity of Neural Representations across the Ventral Stream
    Guclu, Umut
    van Gerven, Marcel A. J.
    JOURNAL OF NEUROSCIENCE, 2015, 35 (27) : 10005 - 10014
  • [45] Understanding Effects of Architecture Design to Invariance and Complexity in Deep Neural Networks
    Kim, Dongha
    Kim, Yongdai
    IEEE ACCESS, 2021, 9 : 9670 - 9681
  • [46] Controlled Dropout: a Different Dropout for Improving Training Speed on Deep Neural Network
    Ko, ByungSoo
    Kim, Han-Gyu
    Choi, Ho-Jin
    2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, : 972 - 977
  • [47] Towards dropout training for convolutional neural networks
    Wu, Haibing
    Gu, Xiaodong
    NEURAL NETWORKS, 2015, 71 : 1 - 10
  • [48] Introduction to Machine Learning, Neural Networks, and Deep Learning
    Choi, Rene Y.
    Coyner, Aaron S.
    Kalpathy-Cramer, Jayashree
    Chiang, Michael F.
    Campbell, J. Peter
    TRANSLATIONAL VISION SCIENCE & TECHNOLOGY, 2020, 9 (02):
  • [49] A General Approach to Dropout in Quantum Neural Networks
    Scala, Francesco
    Ceschini, Andrea
    Panella, Massimo
    Gerace, Dario
    ADVANCED QUANTUM TECHNOLOGIES, 2023,
  • [50] Training Deep Neural Networks with Constrained Learning Parameters
    Date, Prasanna
    Carothers, Christopher D.
    Mitchell, John E.
    Hendler, James A.
    Magdon-Ismail, Malik
    2020 INTERNATIONAL CONFERENCE ON REBOOTING COMPUTING (ICRC 2020), 2020, : 107 - 115