Accelerating neural network training using weight extrapolations

被引:91
作者
Kamarthi, SV [1 ]
Pittner, S [1 ]
机构
[1] Northeastern Univ, Dept Mech Ind & Mfg Engn, Snell Eng Ctr 334, Boston, MA 02115 USA
关键词
multilayer neural networks; backpropagation algorithm; convergence acceleration; extrapolation methods; parameter estimation; linear regression; conjugate gradient method; relative entropy;
D O I
10.1016/S0893-6080(99)00072-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The backpropagation (BP) algorithm for training feedforward neural networks has proven robust even for difficult problems. However, its high performance results are attained at the expense of a long training time to adjust the network parameters, which can be discouraging in many real-world applications. Even on relatively simple problems, standard BP often requires a lengthy training process in which the complete set of training examples is processed hundreds or thousands of times. In this paper, a universal acceleration technique for the BP algorithm based on extrapolation of each individual interconnection weight is presented. This extrapolation procedure is easy to implement and is activated only a few times in between iterations of the conventional BP algorithm. This procedure, unlike earlier acceleration procedures, minimally alters the computational structure of the BP algorithm. The viability of this new approach is demonstrated on three examples. The results suggest that it leads to significant savings in computation time of the standard BP algorithm. Moreover, the solution computed by the proposed approach is always located in close proximity to the one obtained by the conventional BP procedure. Hence, the proposed method provides a real acceleration of the BP algorithm without degrading the usefulness of its solutions. The performance of the new method is also compared with that of the conjugate gradient algorithm, which is an improved and faster version of the BP algorithm. (C) 1999 Elsevier Science Ltd. All rights reserved.
引用
收藏
页码:1285 / 1299
页数:15
相关论文
共 51 条
  • [41] Ryan T. P., 1997, MODERN REGRESSION ME
  • [42] Salehi F, 1998, T ASAE, V41, P253, DOI 10.13031/2013.17144
  • [43] Salomon R, 1996, NEURAL NETWORKS, V9, P589, DOI 10.1016/0893-6080(95)00144-1
  • [44] Sejnowski T. J., 1987, Complex Systems, V1, P145
  • [45] SILVA FM, 1990, LECT NOTES COMPUT SC, V412, P110
  • [46] Asymptotic Convergence of Backpropagation
    Tesauro, Gerald
    He, Yu
    Ahmad, Subutai
    [J]. NEURAL COMPUTATION, 1989, 1 (03) : 382 - 391
  • [47] TOWSEY M, 1995, P IEEE INT C NEUR NE, P373
  • [48] Vetterling W. T, 2002, NUMERICAL RECIPES C
  • [49] A METHOD FOR SELF-DETERMINATION OF ADAPTIVE LEARNING RATES IN BACK PROPAGATION
    WEIR, MK
    [J]. NEURAL NETWORKS, 1991, 4 (03) : 371 - 379
  • [50] Yamada K., 1997, Systems and Computers in Japan, V28, P48, DOI 10.1002/(SICI)1520-684X(199704)28:4<48::AID-SCJ5>3.0.CO