Reconsideration to pruning and regularization for complexity optimization in neural networks

被引:0
作者
Park, H [1 ]
Lee, H [1 ]
机构
[1] RIKEN, Brain Sci Inst, Wako, Saitama 35101, Japan
来源
ICONIP'02: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING: COMPUTATIONAL INTELLIGENCE FOR THE E-AGE | 2002年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The ultimate purpose of neural network design is to find an optimal network that can give good generalization performance with compact structure. To achieve this, it is necessary to control complexities of networks so as to avoid its overfitting to noisy learning data. The most popular methods for complexity control are the pruning method and the regularization method. Even though there have been many variations in the methods, the peculiar properties of each method compared to others has not been so clear. In this paper, we reconsider the pruning strategy from geometrical and statistical viewpoint, and show that the natural pruning method is in accordance with the geometrical and statistical intuition in choosing connections to be pruned. In addition, we also suggest that the regularization method should be used in combination with natural pruning in order to improve the optimization performance. We also show some experimental results supporting our suggestions.
引用
收藏
页码:1649 / 1653
页数:5
相关论文
共 11 条
[1]   Natural gradient works efficiently in learning [J].
Amari, S .
NEURAL COMPUTATION, 1998, 10 (02) :251-276
[2]  
AMARI S, 2001, INFORMATION GEOMETRY
[3]  
Bishop C. M., 1995, NEURAL NETWORKS PATT
[4]  
Hassibi B., 1993, ADV NEURAL INFORM PR, P164, DOI DOI 10.5555/645753.668069
[5]  
HESKES T, 2000, NEURAL COMPUT, V12, P1037
[6]   Neural classifier construction using regularization, pruning and test error estimation [J].
Hintz-Madsen, M ;
Hansen, LK ;
Larsen, J ;
Pedersen, MW ;
Larsen, M .
NEURAL NETWORKS, 1998, 11 (09) :1659-1670
[7]  
LeCun Y., 1990, Advances in neural information processing systems, P598
[8]   Adaptive natural gradient learning algorithms for various stochastic models [J].
Park, H ;
Amari, SI ;
Fukumizu, K .
NEURAL NETWORKS, 2000, 13 (07) :755-764
[9]  
Pedersen MW, 1996, ADV NEUR IN, V8, P521
[10]  
PRECHELT L, 1994, 2194 U KARLSR FAK IN