Deep network in network

被引:22
作者
Alaeddine, Hmidi [1 ]
Jihene, Malek [1 ,2 ]
机构
[1] Monastir Univ, Fac Sci Monastir, Lab Elect & Microelect, LR99ES30, Monastir 5000, Tunisia
[2] Sousse Univ, Higher Inst Appl Sci & Technol Sousse, Sousse 4000, Tunisia
关键词
Exponential linear unit (ELU); Convolutional neural networks (CNNs); Deep MLPconv; Image recognition; Network in Network (NiN); NEURAL-NETWORKS; LEARNING DEEP; REPRESENTATION;
D O I
10.1007/s00521-020-05008-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The different CNN models use many layers that typically include a stack of linear convolution layers combined with pooling and normalization layers to extract the characteristics of the images. Unlike these models, and instead of using a linear filter for convolution, the network in network (NiN) model uses a multilayer perception (MLP), a nonlinear function, to replace the linear filter. This article presents a new deep network in network (DNIN) model based on the NiN structure, NiN drag a universal approximator, (MLP) with rectified linear unit (ReLU) to improve classification performance. The use of MLP leads to an increase in the density of the connection. This makes learning more difficult and time learning slower. In this article, instead of ReLU, we use the linear exponential unit (eLU) to solve the vanishing gradient problem that can occur when using ReLU and to speed up the learning process. In addition, a reduction in the convolution filters size by increasing the depth is used in order to reduce the number of parameters. Finally, a batch normalization layer is applied to reduce the saturation of the eLUs and the dropout layer is applied to avoid overfitting. The experimental results on the CIFAR-10 database show that the DNIN can reduce the complexity of implementation due to the reduction in the adjustable parameters. Also the reduction in the filters size shows an improvement in the recognition accuracy of the model.
引用
收藏
页码:1453 / 1465
页数:13
相关论文
共 48 条
[1]  
Agostinelli F., 2014, arXiv preprint arXiv:1412.6830
[2]  
[Anonymous], 2015, Batch-normalized maxout network in network
[3]  
[Anonymous], 2013, JMLR Workshop and Conference Proceedings
[4]  
[Anonymous], 2015, PREPRINT
[5]  
[Anonymous], 2014, Overfeat: integrated recognition, localization and detection using convolutional networks
[6]  
[Anonymous], 2014, CORR
[7]   Learning Deep Architectures for AI [J].
Bengio, Yoshua .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2009, 2 (01) :1-127
[8]   Pedestrian Detection Inspired by Appearance Constancy and Shape Symmetry [J].
Cao, Jiale ;
Pang, Yanwei ;
Li, Xuelong .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (12) :5538-5551
[9]   Evaluation of maxout activations in deep learning across several big data domains [J].
Castaneda, Gabriel ;
Morris, Paul ;
Khoshgoftaar, Taghi M. .
JOURNAL OF BIG DATA, 2019, 6 (01)
[10]  
CHAN T, 2014, CORR