Selective Dropout for Deep Neural Networks

被引:5
|
作者
Barrow, Erik [1 ]
Eastwood, Mark [1 ]
Jayne, Chrisina [2 ]
机构
[1] Coventry Univ, Coventry, W Midlands, England
[2] Robert Gordon Univ, Aberdeen, Scotland
来源
NEURAL INFORMATION PROCESSING, ICONIP 2016, PT III | 2016年 / 9949卷
关键词
MNIST; Artificial neural network; Deep learning; Dropout network; Non-random dropout; Selective dropout;
D O I
10.1007/978-3-319-46675-0_57
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dropout has been proven to be an effective method for reducing overfitting in deep artificial neural networks. We present 3 new alternative methods for performing dropout on a deep neural network which improves the effectiveness of the dropout method over the same training period. These methods select neurons to be dropped through statistical values calculated using a neurons change in weight, the average size of a neuron's weights, and the output variance of a neuron. We found that increasing the probability of dropping neurons with smaller values of these statistics and decreasing the probability of those with larger statistics gave an improved result in training over 10,000 epochs. The most effective of these was found to be the Output Variance method, giving an average improvement of 1.17% accuracy over traditional dropout methods.
引用
收藏
页码:519 / 528
页数:10
相关论文
共 50 条
  • [41] Selection dynamics for deep neural networks
    Liu, Hailiang
    Markowich, Peter
    JOURNAL OF DIFFERENTIAL EQUATIONS, 2020, 269 (12) : 11540 - 11574
  • [42] IMPLICIT SALIENCY IN DEEP NEURAL NETWORKS
    Sun, Yutong
    Prabhushankar, Mohit
    AlRegib, Ghassan
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 2915 - 2919
  • [43] MULTILINGUAL TRAINING OF DEEP NEURAL NETWORKS
    Ghoshal, Arnab
    Swietojanski, Pawel
    Renals, Steve
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 7319 - 7323
  • [44] Hyperbolic Deep Neural Networks: A Survey
    Peng, Wei
    Varanka, Tuomas
    Mostafa, Abdelrahman
    Shi, Henglin
    Zhao, Guoying
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (12) : 10023 - 10044
  • [45] Lossless Compression of Deep Neural Networks
    Serra, Thiago
    Kumar, Abhinav
    Ramalingam, Srikumar
    INTEGRATION OF CONSTRAINT PROGRAMMING, ARTIFICIAL INTELLIGENCE, AND OPERATIONS RESEARCH, CPAIOR 2020, 2020, 12296 : 417 - 430
  • [46] Continuously Constructive Deep Neural Networks
    Irsoy, Ozan
    Alpaydin, Ethem
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (04) : 1124 - 1133
  • [47] Visual Genealogy of Deep Neural Networks
    Wang, Qianwen
    Yuan, Jun
    Chen, Shuxin
    Su, Hang
    Qu, Huamin
    Liu, Shixia
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2020, 26 (11) : 3340 - 3352
  • [48] Fast learning in Deep Neural Networks
    Chandra, B.
    Sharma, Rajesh K.
    NEUROCOMPUTING, 2016, 171 : 1205 - 1215
  • [49] Transfer Entropy in Deep Neural Networks
    Andonie, R.
    Cataron, A.
    Moldovan, A.
    INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2025, 20 (01)
  • [50] A survey on the applications of Deep Neural Networks
    Latha, R. S.
    Sreekanth, G. R. R.
    Suganthe, R. C.
    Selvaraj, R. Esakki
    2021 INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND INFORMATICS (ICCCI), 2021,