Dropout Rademacher complexity of deep neural networks

被引:1
|
作者
Wei GAO [1 ,2 ]
Zhi-Hua ZHOU [1 ,2 ]
机构
[1] National Key Laboratory for Novel Software Technology, Nanjing University
[2] Collaborative Innovation Center of Novel Software Technology and Industrialization,Nanjing University
基金
中国国家自然科学基金;
关键词
artificial intelligence; machine learning; deep learning; dropout; Rademacher complexity;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Great successes of deep neural networks have been witnessed in various real applications. Many algorithmic and implementation techniques have been developed; however, theoretical understanding of many aspects of deep neural networks is far from clear. A particular interesting issue is the usefulness of dropout,which was motivated from the intuition of preventing complex co-adaptation of feature detectors. In this paper,we study the Rademacher complexity of different types of dropouts, and our theoretical results disclose that for shallow neural networks(with one or none hidden layer) dropout is able to reduce the Rademacher complexity in polynomial, whereas for deep neural networks it can amazingly lead to an exponential reduction.
引用
收藏
页码:173 / 184
页数:12
相关论文
共 50 条
  • [1] Dropout Rademacher complexity of deep neural networks
    Gao, Wei
    Zhou, Zhi-Hua
    SCIENCE CHINA-INFORMATION SCIENCES, 2016, 59 (07)
  • [2] Dropout Rademacher complexity of deep neural networks
    Wei Gao
    Zhi-Hua Zhou
    Science China Information Sciences, 2016, 59
  • [3] Rademacher dropout: An adaptive dropout for deep neural network via optimizing generalization gap
    Wang, Haotian
    Yang, Wenjing
    Zhao, Zhenyu
    Luo, Tingjin
    Wang, Ji
    Tang, Yuhua
    NEUROCOMPUTING, 2019, 357 : 177 - 187
  • [4] Selective Dropout for Deep Neural Networks
    Barrow, Erik
    Eastwood, Mark
    Jayne, Chrisina
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT III, 2016, 9949 : 519 - 528
  • [5] Deep Learning Convolutional Neural Networks with Dropout - a Parallel Approach
    Shen, Jingyi
    Shafiq, M. Omair
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, : 572 - 577
  • [6] Regularization of deep neural networks with spectral dropout
    Khan, Salman H.
    Hayat, Munawar
    Porikli, Fatih
    NEURAL NETWORKS, 2019, 110 : 82 - 90
  • [7] IMPROVING DEEP NEURAL NETWORKS BY USING SPARSE DROPOUT STRATEGY
    Zheng, Hao
    Chen, Mingming
    Liu, Wenju
    Yang, Zhanlei
    Liang, Shan
    2014 IEEE CHINA SUMMIT & INTERNATIONAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (CHINASIP), 2014, : 21 - 26
  • [8] Bounding the Rademacher complexity of Fourier neural operators
    Kim, Taeyoung
    Kang, Myungjoo
    MACHINE LEARNING, 2024, 113 (05) : 2467 - 2498
  • [9] Bounding the Rademacher complexity of Fourier neural operators
    Taeyoung Kim
    Myungjoo Kang
    Machine Learning, 2024, 113 : 2467 - 2498
  • [10] Dropout with Tabu Strategy for Regularizing Deep Neural Networks
    Ma, Zongjie
    Sattar, Abdul
    Zhou, Jun
    Chen, Qingliang
    Su, Kaile
    COMPUTER JOURNAL, 2020, 63 (07) : 1031 - 1038