Robust Adaptive Gradient-Descent Training Algorithm for Recurrent Neural Networks in Discrete Time Domain

被引:68
|
作者
Song, Qing [1 ]
Wu, Yilei [1 ]
Soh, Yeng Chai [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2008年 / 19卷 / 11期
关键词
Conic sector condition; hybrid training; real-time recurrent learning (RTRL); recurrent neural network (RNN); weight convergence;
D O I
10.1109/TNN.2008.2001923
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For a recurrent neural network (RNN), its transient response is a critical issue, especially for real-time signal processing applications. The conventional RNN training algorithms, such as backpropagation through time (BPTT) and real-time recurrent learning (RTRL), have not adequately addressed this problem because they suffer from low convergence speed. While increasing the learning rate may help to improve the performance of the RNN, it can result in unstable training in terms of weight divergence. Therefore, an optimal tradeoff between RNN training speed and weight convergence is desired. In this paper, a robust adaptive gradient-descent (RAGD) training algorithm of RNN is developed based on a novel RNN hybrid training concept. It switches the training patterns between standard real-time online backpropagation (BP) and RTRL according to the derived convergence and stability conditions. The weight convergence and L-2-stability of the algorithm are derived via the conic sector theorem. The optimized adaptive learning maximizes the training speed of the RNN for each weight update without violating the stability and convergence criteria. Computer simulations are carried out to demonstrate the applicability of the theoretical results.
引用
收藏
页码:1841 / 1853
页数:13
相关论文
共 50 条
  • [41] A Robust Evolutionary Algorithm for Training Neural Networks
    Jinn-Moon Yang
    Cheng-Yan Kao
    Neural Computing & Applications, 2001, 10 : 214 - 230
  • [42] Training Neural Networks Using Predictor-Corrector Gradient Descent
    Nesky, Amy
    Stout, Quentin F.
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT III, 2018, 11141 : 62 - 72
  • [43] Training Morphological Neural Networks with Gradient Descent: Some Theoretical Insights
    Blusseau, Samy
    DISCRETE GEOMETRY AND MATHEMATICAL MORPHOLOGY, DGMM 2024, 2024, 14605 : 229 - 241
  • [44] ABNGrad: adaptive step size gradient descent for optimizing neural networks
    Wenhan Jiang
    Yuqing Liang
    Zhixia Jiang
    Dongpo Xu
    Linhua Zhou
    Applied Intelligence, 2024, 54 : 2361 - 2378
  • [45] A conjugate gradient learning algorithm for recurrent neural networks
    Chang, WF
    Mak, MW
    NEUROCOMPUTING, 1999, 24 (1-3) : 173 - 189
  • [46] ABNGrad: adaptive step size gradient descent for optimizing neural networks
    Jiang, Wenhan
    Liang, Yuqing
    Jiang, Zhixia
    Xu, Dongpo
    Zhou, Linhua
    APPLIED INTELLIGENCE, 2024, 54 (03) : 2361 - 2378
  • [47] Damped Newton Stochastic Gradient Descent Method for Neural Networks Training
    Zhou, Jingcheng
    Wei, Wei
    Zhang, Ruizhi
    Zheng, Zhiming
    MATHEMATICS, 2021, 9 (13)
  • [48] Discrete-Time ε-Adaptive Dynamic Programming Algorithm Using Neural Networks
    Jin, Ning
    Liu, Derong
    PROCEEDINGS OF THE 2008 IEEE INTERNATIONAL SYMPOSIUM ON INTELLIGENT CONTROL, 2008, : 114 - 119
  • [49] A hybrid model of heuristic algorithm and gradient descent to optimize neural networks
    Mirkhan, Amer
    Celebi, Numan
    BULLETIN OF THE POLISH ACADEMY OF SCIENCES-TECHNICAL SCIENCES, 2023, 71 (06)
  • [50] AN ADAPTIVE TRAINING ALGORITHM FOR BACKPROPAGATION NEURAL NETWORKS
    HSIN, HC
    LI, CC
    SUN, MG
    SCLABASSI, RJ
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1995, 25 (03): : 512 - 514