A New Weight Initialization Method for Sigmoidal Feedforward Artificial Neural Networks

被引:0
作者
Sodhi, Sartaj Singh [1 ]
Chandra, Pravin [1 ]
Tanwar, Sharad [2 ]
机构
[1] Guru Gobind Singh Indraprastha Univ, Univ Sch Inthnnat & Commun Technol, Sect 16C, New Delhi 110078, India
[2] Deloitte Consulting India Private Ltd, Gurgaon 122015, Haryana, India
来源
PROCEEDINGS OF THE 2014 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2014年
关键词
IMPROVING TRAINING SPEED; BACKPROPAGATION; OPTIMIZATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Initial weight choice has been recognized to be an important aspect of the training methodology for sigmoidal feedforward neural networks. In this paper, a new mechanism for weight initialization is proposed. The mechanism distributes the initial input to output weights in a manner that all weights (including thresholds) leading into a hidden layer are uniformly distributed in a region and the center of the region from which the weights are sampled are such that no region overlaps for two distinct hidden nodes. The proposed method is compared against random weight initialization routines on five function approximation tasks using the Resilient Backpropagation (RPROP) algorithm for training. The proposed method is shown to lead to about twice as fast convergence to a pre-specified goal for training as compared to any of the random weight initialization methods. Moreover, it is shown that at least for these problems the networks reach a deeper minima of the error functional during training and generalizes better than the networks trained whose weights were initialized by random weight initialization methods.
引用
收藏
页码:291 / 298
页数:8
相关论文
共 36 条
[1]  
[Anonymous], 1998, Learning from data-concepts, theory and methods
[2]   THE PI METHOD FOR ESTIMATING MULTIVARIATE FUNCTIONS FROM NOISY DATA [J].
BREIMAN, L .
TECHNOMETRICS, 1991, 33 (02) :125-143
[3]   Sigmoidal Function Classes for Feedforward Artificial Neural Networks [J].
Pravin Chandra .
Neural Processing Letters, 2003, 18 (3) :205-215
[4]  
CHEN CL, 1991, IEEE IJCNN, P2063, DOI 10.1109/IJCNN.1991.170691
[5]   Comparison of adaptive methods for function estimation from samples [J].
Cherkassky, V ;
Gehring, D .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1996, 7 (04) :969-984
[6]  
Cybenko G., 1989, Mathematics of Control, Signals, and Systems, V2, P303, DOI 10.1007/BF02551274
[7]   Feedforward neural network initialization: an evolutionary approach [J].
de Castro, LN ;
Iyoda, EM ;
Von Zuben, FJ ;
Gudwin, R .
VTH BRAZILIAN SYMPOSIUM ON NEURAL NETWORKS, PROCEEDINGS, 1998, :43-48
[8]   STATISTICALLY CONTROLLED ACTIVATION WEIGHT INITIALIZATION (SCAWI) [J].
DRAGO, GP ;
RIDELLA, S .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (04) :627-631
[9]  
Duch W., 1999, Neural Computing Surveys, V2
[10]  
Fahlman Scott E., 1988, An Empirical Study of Learning Speed in Back-Propagation Networks