A sequential learning approach for single hidden layer neural networks

被引:91
作者
Zhang, J [1 ]
Morris, AJ [1 ]
机构
[1] Newcastle Univ, Dept Chem & Proc Engn, Ctr Proc Anal Chemometr & Control, Newcastle Upon Tyne NE1 7RU, Tyne & Wear, England
关键词
neural networks; neural network structures; neural network training; non-linear system modelling;
D O I
10.1016/S0893-6080(97)00111-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A sequential orthogonal approach to the building and training of single hidden layer neural networks is presented in this paper. In the proposed method, hidden neurons are added one at a time. The procedure starts with a single hidden neuron and sequentially increases the number of hidden neurons until the model error is sufficiently small. When adding a neuron, the new information introduced by this neuron is caused by that part of its output vector which is orthogonal to the space spanned by the output vectors of previously added hidden neurons. The classical Gram-Schmidt orthogonalization method is used at each step to form a set of orthogonal bases for the space spanned by the output vectors of hidden neurons. Hidden layer weights are found through optimization while output layer weights are obtained from the least-squares regression. Using the proposed technique, it is possible to determine the necessary number of hidden neurons required. A regularization factor is also incorporated into the sequential orthogonal training algorithm to improve the network generalization capability. An additional advantage of this method is that it can be used to build and train neural networks with mixed types of hidden neurons and thus to develop hybrid models. By using mixed types of neurons, it is found that more accurate neural network models, with a smaller number of hidden neurons than seen in conventional networks, can be developed. The proposed sequential orthogonal training method was successfully applied to three non-linear modelling examples. (C) 1998 Elsevier Science Ltd. All rights reserved.
引用
收藏
页码:65 / 80
页数:16
相关论文
共 33 条
  • [1] NEW LOOK AT STATISTICAL-MODEL IDENTIFICATION
    AKAIKE, H
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1974, AC19 (06) : 716 - 723
  • [2] BALLARD DH, 1988, VISION BRAIN COOPERA, P563
  • [3] USE OF NEURAL NETS FOR DYNAMIC MODELING AND CONTROL OF CHEMICAL PROCESS SYSTEMS
    BHAT, N
    MCAVOY, TJ
    [J]. COMPUTERS & CHEMICAL ENGINEERING, 1990, 14 (4-5) : 573 - 583
  • [4] BILLINGS SA, 1986, INT J CONTROL, V44, P235, DOI 10.1080/00207178608933593
  • [5] Improving the Generalization Properties of Radial Basis Function Neural Networks
    Bishop, Chris
    [J]. NEURAL COMPUTATION, 1991, 3 (04) : 579 - 588
  • [6] PRACTICAL IDENTIFICATION OF NARMAX MODELS USING RADIAL BASIS FUNCTIONS
    CHEN, S
    BILLINGS, SA
    COWAN, CFN
    GRANT, PM
    [J]. INTERNATIONAL JOURNAL OF CONTROL, 1990, 52 (06) : 1327 - 1350
  • [7] Regularized orthogonal least squares algorithm for constructing radial basis function networks
    Chen, S
    Chng, ES
    Alkadhimi, K
    [J]. INTERNATIONAL JOURNAL OF CONTROL, 1996, 64 (05) : 829 - 837
  • [8] ORTHOGONAL LEAST-SQUARES LEARNING ALGORITHM FOR RADIAL BASIS FUNCTION NETWORKS
    CHEN, S
    COWAN, CFN
    GRANT, PM
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1991, 2 (02): : 302 - 309
  • [9] NONLINEAR-SYSTEM IDENTIFICATION USING NEURAL NETWORKS
    CHEN, S
    BILLINGS, SA
    GRANT, PM
    [J]. INTERNATIONAL JOURNAL OF CONTROL, 1990, 51 (06) : 1191 - 1214
  • [10] CHENTOUF R, 1996, P 4 EUR S ART NEUR N, P129