A sequential learning approach for single hidden layer neural networks

被引:91
作者
Zhang, J [1 ]
Morris, AJ [1 ]
机构
[1] Newcastle Univ, Dept Chem & Proc Engn, Ctr Proc Anal Chemometr & Control, Newcastle Upon Tyne NE1 7RU, Tyne & Wear, England
关键词
neural networks; neural network structures; neural network training; non-linear system modelling;
D O I
10.1016/S0893-6080(97)00111-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A sequential orthogonal approach to the building and training of single hidden layer neural networks is presented in this paper. In the proposed method, hidden neurons are added one at a time. The procedure starts with a single hidden neuron and sequentially increases the number of hidden neurons until the model error is sufficiently small. When adding a neuron, the new information introduced by this neuron is caused by that part of its output vector which is orthogonal to the space spanned by the output vectors of previously added hidden neurons. The classical Gram-Schmidt orthogonalization method is used at each step to form a set of orthogonal bases for the space spanned by the output vectors of hidden neurons. Hidden layer weights are found through optimization while output layer weights are obtained from the least-squares regression. Using the proposed technique, it is possible to determine the necessary number of hidden neurons required. A regularization factor is also incorporated into the sequential orthogonal training algorithm to improve the network generalization capability. An additional advantage of this method is that it can be used to build and train neural networks with mixed types of hidden neurons and thus to develop hybrid models. By using mixed types of neurons, it is found that more accurate neural network models, with a smaller number of hidden neurons than seen in conventional networks, can be developed. The proposed sequential orthogonal training method was successfully applied to three non-linear modelling examples. (C) 1998 Elsevier Science Ltd. All rights reserved.
引用
收藏
页码:65 / 80
页数:16
相关论文
共 33 条
[11]  
Cybenko G., 1989, Mathematics of Control, Signals, and Systems, V2, P303, DOI 10.1007/BF02551274
[12]  
Fahlman S., 1990, ADV NEURAL INFORMATI, V2, P524
[13]   PROJECTION PURSUIT REGRESSION [J].
FRIEDMAN, JH ;
STUETZLE, W .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1981, 76 (376) :817-823
[14]   NONLINEAR MULTIVARIATE CALIBRATION USING PRINCIPAL COMPONENTS REGRESSION AND ARTIFICIAL NEURAL NETWORKS [J].
GEMPERLINE, PJ ;
LONG, JR ;
GREGORIOU, VG .
ANALYTICAL CHEMISTRY, 1991, 63 (20) :2313-2323
[15]   NETWORKS AND THE BEST APPROXIMATION PROPERTY [J].
GIROSI, F ;
POGGIO, T .
BIOLOGICAL CYBERNETICS, 1990, 63 (03) :169-176
[16]  
Haesloop D., 1990, Proceedings of the 1990 American Control Conference (IEEE Cat. No.90CH2896-9), P2460
[17]  
Hofland A. G., 1992, Proceedings of the 1992 American Control Conference (IEEE Cat. No.92CH3072-6), P480
[18]  
Holcomb T., 1991, Proceedings of the 1991 American Control Conference (IEEE Cat. No. 91CH2939-7), P2331
[19]   PLS NEURAL NETWORKS [J].
HOLCOMB, TR ;
MORARI, M .
COMPUTERS & CHEMICAL ENGINEERING, 1992, 16 (04) :393-411
[20]  
HWANG JN, 1994, IEEE T NEURAL NETWOR, V5, P342, DOI 10.1109/72.286906