Group L1/2 Regularization for Pruning Hidden Layer Nodes of Feedforward Neural Networks

被引:16
作者
Alemu, Habtamu Zegeye [1 ]
Zhao, Junhong [1 ]
Li, Feng [1 ]
Wu, Wei [1 ]
机构
[1] Dalian Univ Technol, Sch Math Sci, Dalian 116024, Peoples R China
关键词
Feedforward neural networks; pruning hidden layer nodes and weights; group L-1(/2); smooth group L-1/2; group lasso; convergence; SMOOTHING L-1/2 REGULARIZATION; GRADIENT LEARNING ALGORITHM; CONVERGENCE; REGRESSION; SELECTION; PENALTY;
D O I
10.1109/ACCESS.2018.2890740
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A group L-1(/2) regularization term is defined and introduced into the conventional error function for pruning the hidden layer nodes of feedforward neural networks. This group L-1(/2) regularization method (GL(1/2)) can prune not only the redundant hidden nodes but also the redundant weights of the surviving hidden nodes of the neural networks. As a comparison, the popular group lasso regularization (GL(2)) can prune the redundant hidden nodes, but cannot prune any redundant weights of the surviving hidden nodes, of the neural networks. A disadvantage of the GL(1/2) is that it involves a non-smooth absolute value function, which causes oscillation in the numerical computation and difficulty in the convergence analysis. As a remedy, the absolute value function is approximated by a smooth function, resulting in a smooth group L-1(/2) regularization method (SGL(1/2)). Numerical simulations on a few benchmark data sets show that, compared with GL(2), SGL(1/2) can achieve better accuracy and remove more redundant nodes and weights of the surviving hidden nodes. A convergence theorem is also proved for SGL(1/2).
引用
收藏
页码:9540 / 9557
页数:18
相关论文
共 44 条
[1]   Feedforward Neural Networks with a Hidden Layer Regularization Method [J].
Alemu, Habtamu Zegeye ;
Wu, Wei ;
Zhao, Junhong .
SYMMETRY-BASEL, 2018, 10 (10)
[2]  
ALVAREZ JM, 2016, ADV NEURAL INFORM PR, P2270
[3]  
[Anonymous], 2013, INTRO STAT LEARNING
[4]  
[Anonymous], 2006, J ROYAL STAT SOC B
[5]   Pruning algorithms of neural networks - a comparative study [J].
Augasta, M. Gethsiyal ;
Kathirvalavakumar, T. .
OPEN COMPUTER SCIENCE, 2013, 3 (03) :105-115
[6]  
Bartlett PL, 1997, ADV NEUR IN, V9, P134
[7]   A ROBUST BACK-PROPAGATION LEARNING ALGORITHM FOR FUNCTION APPROXIMATION [J].
CHEN, DS ;
JAIN, RC .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (03) :467-479
[8]  
Cui ZH, 2004, I C CONT AUTOMAT ROB, P238
[9]  
Denil Misha, 2013, NIPS, P2148
[10]  
Dheeru D., 2017, TECH REP