Dropout Rademacher complexity of deep neural networks

被引:1
|
作者
Wei GAO [1 ,2 ]
Zhi-Hua ZHOU [1 ,2 ]
机构
[1] National Key Laboratory for Novel Software Technology, Nanjing University
[2] Collaborative Innovation Center of Novel Software Technology and Industrialization,Nanjing University
基金
中国国家自然科学基金;
关键词
artificial intelligence; machine learning; deep learning; dropout; Rademacher complexity;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Great successes of deep neural networks have been witnessed in various real applications. Many algorithmic and implementation techniques have been developed; however, theoretical understanding of many aspects of deep neural networks is far from clear. A particular interesting issue is the usefulness of dropout,which was motivated from the intuition of preventing complex co-adaptation of feature detectors. In this paper,we study the Rademacher complexity of different types of dropouts, and our theoretical results disclose that for shallow neural networks(with one or none hidden layer) dropout is able to reduce the Rademacher complexity in polynomial, whereas for deep neural networks it can amazingly lead to an exponential reduction.
引用
收藏
页码:173 / 184
页数:12
相关论文
共 50 条
  • [31] Impact of deep learning-based dropout on shallow neural networks applied to stream temperature modelling
    Piotrowski, Adam P.
    Napiorkowski, Jaroslaw J.
    Piotrowska, Agnieszka E.
    EARTH-SCIENCE REVIEWS, 2020, 201
  • [32] Rademacher Complexity of Neural ODEs via Chen-Fliess Series
    Hanson, Joshua
    Raginsky, Maxim
    6TH ANNUAL LEARNING FOR DYNAMICS & CONTROL CONFERENCE, 2024, 242 : 758 - 769
  • [33] A Deep Connection Between the Vapnik-Chervonenkis Entropy and the Rademacher Complexity
    Anguita, Davide
    Ghio, Alessandro
    Oneto, Luca
    Ridella, Sandro
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (12) : 2202 - 2211
  • [34] A Dropout Distribution Model on Deep Networks
    Li, Fengqi
    Yang, Helin
    EIGHTH INTERNATIONAL CONFERENCE ON DIGITAL IMAGE PROCESSING (ICDIP 2016), 2016, 10033
  • [35] Universal Approximation in Dropout Neural Networks
    Manita, Oxana A.
    Peletier, Mark A.
    Portegies, Jacobus W.
    Sanders, Jaron
    Senen-Cerda, Albert
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [36] Surprising properties of dropout in deep networks
    Helmbold, David P.
    Long, Philip M.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2018, 18
  • [37] Anomaly Detection Approach Based on Deep Neural Network and Dropout
    Hussien, Zaid Khalaf
    Dhannoon, Ban N.
    BAGHDAD SCIENCE JOURNAL, 2020, 17 (02) : 701 - 709
  • [38] Shift Quality Classifier Using Deep Neural Networks on Small Data with Dropout and Semi-Supervised Learning
    Kawakami, Takefumi
    Ide, Takanori
    Hoki, Kunihito
    Muramatsu, Masakazu
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2023, E106D (12) : 2078 - 2084
  • [39] Understanding Dropout for Graph Neural Networks
    Shu, Juan
    Xi, Bowei
    Li, Yu
    Wu, Fan
    Kamhoua, Charles
    Ma, Jianzhu
    COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2022, WWW 2022 COMPANION, 2022, : 1128 - 1138
  • [40] Random image frequency aggregation dropout in image classification for deep convolutional neural networks
    Nam, Ju-Hyeon
    Lee, Sang-Chul
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 232