Robust Adaptive Gradient-Descent Training Algorithm for Recurrent Neural Networks in Discrete Time Domain

被引:68
|
作者
Song, Qing [1 ]
Wu, Yilei [1 ]
Soh, Yeng Chai [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2008年 / 19卷 / 11期
关键词
Conic sector condition; hybrid training; real-time recurrent learning (RTRL); recurrent neural network (RNN); weight convergence;
D O I
10.1109/TNN.2008.2001923
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For a recurrent neural network (RNN), its transient response is a critical issue, especially for real-time signal processing applications. The conventional RNN training algorithms, such as backpropagation through time (BPTT) and real-time recurrent learning (RTRL), have not adequately addressed this problem because they suffer from low convergence speed. While increasing the learning rate may help to improve the performance of the RNN, it can result in unstable training in terms of weight divergence. Therefore, an optimal tradeoff between RNN training speed and weight convergence is desired. In this paper, a robust adaptive gradient-descent (RAGD) training algorithm of RNN is developed based on a novel RNN hybrid training concept. It switches the training patterns between standard real-time online backpropagation (BP) and RTRL according to the derived convergence and stability conditions. The weight convergence and L-2-stability of the algorithm are derived via the conic sector theorem. The optimized adaptive learning maximizes the training speed of the RNN for each weight update without violating the stability and convergence criteria. Computer simulations are carried out to demonstrate the applicability of the theoretical results.
引用
收藏
页码:1841 / 1853
页数:13
相关论文
共 50 条
  • [1] Recurrent neural tracking control based on multivariable robust adaptive gradient-descent training algorithm
    Xu, Zhao
    Song, Qing
    Wang, Danwei
    NEURAL COMPUTING & APPLICATIONS, 2012, 21 (07): : 1745 - 1755
  • [2] Recurrent neural tracking control based on multivariable robust adaptive gradient-descent training algorithm
    Zhao Xu
    Qing Song
    Danwei Wang
    Neural Computing and Applications, 2012, 21 : 1745 - 1755
  • [3] Differentiable Spike: Rethinking Gradient-Descent for Training Spiking Neural Networks
    Li, Yuhang
    Guo, Yufei
    Zhang, Shanghang
    Deng, Shikuang
    Hai, Yongqing
    Gu, Shi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [4] Adaptive nonmonotone conjugate gradient training algorithm for recurrent neural networks
    Peng, Chun-Cheng
    Magoulas, George D.
    19TH IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, VOL II, PROCEEDINGS, 2007, : 374 - 381
  • [5] Convergence of Gradient Descent Algorithm for Diagonal Recurrent Neural Networks
    Xu, Dongpo
    Li, Zhengxue
    Wu, Wei
    Ding, Xiaoshuai
    Qu, Di
    2007 SECOND INTERNATIONAL CONFERENCE ON BIO-INSPIRED COMPUTING: THEORIES AND APPLICATIONS, 2007, : 29 - 31
  • [6] ADVANCED ADAPTIVE NONMONOTONE CONJUGATE GRADIENT TRAINING ALGORITHM FOR RECURRENT NEURAL NETWORKS
    Peng, Chun-Cheng
    Magoulas, George D.
    INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, 2008, 17 (05) : 963 - 984
  • [7] Time delay learning by gradient descent in Recurrent Neural Networks
    Boné, R
    Cardot, H
    ARTIFICIAL NEURAL NETWORKS: FORMAL MODELS AND THEIR APPLICATIONS - ICANN 2005, PT 2, PROCEEDINGS, 2005, 3697 : 175 - 180
  • [8] Training Neural Networks by Time-Fractional Gradient Descent
    Xie, Jingyi
    Li, Sirui
    AXIOMS, 2022, 11 (10)
  • [9] Exponential convergence of a gradient descent algorithm for a class of recurrent neural networks
    Bartlett, P
    Dasgupta, S
    38TH MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS, PROCEEDINGS, VOLS 1 AND 2, 1996, : 497 - 500
  • [10] Convergence of gradient descent algorithm with penalty term for recurrent neural networks
    Ding, Xiaoshuai
    Wang, Kuaini
    International Journal of Multimedia and Ubiquitous Engineering, 2014, 9 (09): : 151 - 158