Improving neural network training solutions using regularisation

被引:62
作者
Mc Loone, S [1 ]
Irwin, G [1 ]
机构
[1] Queens Univ Belfast, Sch Elect & Elect Engn, Intelligent Syst & Control Res Grp, Belfast BT9 5AH, Antrim, North Ireland
关键词
neural networks; training; hybrid optimisation; regularisation;
D O I
10.1016/S0925-2312(00)00314-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes the application of regularisation to the training of feedforward neural networks, as a means of improving the quality of solutions obtained. The basic principles of regularisation theory are outlined for both linear and nonlinear training and then extended to cover a new hybrid training algorithm for feedforward neural networks recently proposed by the authors. The concept of functional regularisation is also introduced and discussed in relation to MLP and RBF networks. The tendency for the hybrid training algorithm and many linear optimisation strategies to generate large magnitude weight solutions when applied to ill-conditioned neural paradigms is illustrated graphically and reasoned analytically. While such weight solutions do not generally result in poor fits, it is argued that they could be subject to numerical instability and are therefore undesirable. Using an illustrative example it is shown that, as well as being beneficial from a generalisation perspective, regularisation also provides a means for controlling the magnitude of solutions. (C) 2001 Elsevier Science B.V. All rights reserved.
引用
收藏
页码:71 / 90
页数:20
相关论文
共 33 条
[1]  
[Anonymous], 1986, PARALLEL DISTRIBUTED
[2]  
[Anonymous], 1989, ADV NEURAL INFORM PR
[3]   1ST-ORDER AND 2ND-ORDER METHODS FOR LEARNING - BETWEEN STEEPEST DESCENT AND NEWTON METHOD [J].
BATTITI, R .
NEURAL COMPUTATION, 1992, 4 (02) :141-166
[4]  
Battiti R., 1990, INCC 90 PAR INT NEUR, V2, P757
[5]   Improving the Generalization Properties of Radial Basis Function Neural Networks [J].
Bishop, Chris .
NEURAL COMPUTATION, 1991, 3 (04) :579-588
[6]  
BISHOP CM, 1990, P INT NEUR NETW C, V90, P749
[7]  
BOSSLEY KM, 1997, THESIS U SOUTHAMPTON
[8]   HYBRID LEARNING ALGORITHM FOR GAUSSIAN POTENTIAL FUNCTION NETWORKS [J].
CHEN, CL ;
CHEN, WC ;
CHANG, FY .
IEE PROCEEDINGS-D CONTROL THEORY AND APPLICATIONS, 1993, 140 (06) :442-448
[9]   APPROXIMATION CAPABILITY TO FUNCTIONS OF SEVERAL VARIABLES, NONLINEAR FUNCTIONALS, AND OPERATORS BY RADIAL BASIS FUNCTION NEURAL NETWORKS [J].
CHEN, TP ;
CHEN, H .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1995, 6 (04) :904-910
[10]   IMPROVING GENERALIZATION PERFORMANCE USING DOUBLE BACKPROPAGATION [J].
DRUCKER, H ;
LECUN, Y .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (06) :991-997