DropFilterR: A Novel Regularization Method for Learning Convolutional Neural Networks

被引:9
作者
Pan, Hengyue [1 ]
Niu, Xin [1 ]
Li, Rongchun [1 ]
Shen, Siqi [1 ]
Dou, Yong [1 ]
机构
[1] Natl Univ Def Technol, Sch Comp, Changsha, Hunan, Peoples R China
关键词
CNNs; Regularization methods; DropFilterR; DEEP; DROPOUT;
D O I
10.1007/s11063-019-10147-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The past few years have witnessed the fast development of regularization methods for deep learning models such as fully-connected deep neural networks (DNNs) and convolutional neural networks (CNNs). Part of previous methods mainly consider to drop features from input data and hidden layers, such as Dropout, Cutout and DropBlocks. DropConnect select to drop connections between fully-connected layers. By randomly discard some features or connections, the above mentioned methods relieve the overfitting problem and improve the performance of neural networks. In this paper, we proposed a novel regularization methods, namely DropFilterR, for the learning of CNNs. The basic idea of DropFilterR is to relax the rule of weight-sharing in CNNs by randomly drop elements in convolution filters. Specifically, we drop different elements in convolution filters along with their moving on input feature maps. Moreover, we may apply random drop rate to further increase the randomness of the proposed method. Also, we find a suitable way to accelerate the computation for DropFilterR based on theoretical analysis. Experimental results on several widely-used image databases such as MNIST, CIFAR-10 and Pascal VOC 2012 show that using DropFilterR improves performance on image classification tasks.
引用
收藏
页码:1285 / 1298
页数:14
相关论文
共 29 条
[1]  
[Anonymous], COMPUTER VISION ECCV
[2]  
[Anonymous], 2014, Comput. Sci.
[3]  
[Anonymous], ARXIV14124564 CORR
[4]  
[Anonymous], 2013, Advances in Neural Information Processing Systems, DOI DOI 10.48550/ARXIV.1307.1493
[5]  
[Anonymous], 2017, ARXIV170804552 CORR
[6]  
[Anonymous], 2022, IEEE T PATTERN ANAL, DOI [DOI 10.1109/CVPR46437.2021.01135, DOI 10.1109/TPAMI.2019.2932058]
[7]  
[Anonymous], ARXIV181009849
[8]  
[Anonymous], ARXIV181012890 CORR
[9]  
Ba J., 2013, Advances in Neural Information Processing Systems, P3084, DOI DOI 10.1109/ACCESS.2019.2929174
[10]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848