Data Dropout in Arbitrary Basis for Deep Network Regularization

被引:0
作者
Rahmani, Mostafa [1 ]
Atia, George K. [1 ]
机构
[1] Univ Cent Florida, Dept Elect & Comp Engn, Orlando, FL 32816 USA
来源
2018 CONFERENCE RECORD OF 52ND ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS | 2018年
关键词
Deep Learning; Dropout; Regularization; Data Projection; Random Projection;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In training deep networks with high capacity, it is important to ensure that the trained network works well when presented with new inputs outside the training dalaset. Dropout is an effective regularization technique to boost the network's generalization in which a random subset of the elements of the given data and the extracted features are set to zero during the training process. In this paper, a new randomized regularization technique in which we withhold a random part of the data without necessarily turning off the neurons/dataelements is proposed. In the proposed method, of which the conventional dropout is shown to be a special case, random data dropout is performed in an arbitrary basis, hence the designation Generalized Dropout. We also present a framework whereby the proposed technique can be applied efficiently to convolutional neural networks. The presented numerical experiments demonstrate that the proposed technique yields a notable performance gain. Generalized Dropout provides new insight into the idea of dropout, shows that we can achieve different performance gains by using different bases matrices, and opens up a new research question as of how to choose optimal bases matrices that achieve maximal performance gain.
引用
收藏
页码:66 / 70
页数:5
相关论文
共 25 条
[1]  
[Anonymous], PROC CVPR IEEE
[2]  
[Anonymous], 2013, Advances in neural information processing systems
[3]  
[Anonymous], 2013, INT C MACHINE LEARNI
[4]  
[Anonymous], 2013, Advances in Neural Information Processing Systems, DOI DOI 10.48550/ARXIV.1307.1493
[5]  
[Anonymous], 2016, BMVC
[6]  
[Anonymous], 2017, IEEE C COMPUTER VISI, DOI DOI 10.1109/CVPR.2017.243
[7]  
[Anonymous], 2016, ARXIV161106148
[8]  
[Anonymous], 2013, ARXIV PREPRINT ARXIV
[9]  
[Anonymous], 2013, ICML
[10]  
Bahdanau D., 2015, Neural machine translation