Robust Adaptive Gradient-Descent Training Algorithm for Recurrent Neural Networks in Discrete Time Domain

被引:68
|
作者
Song, Qing [1 ]
Wu, Yilei [1 ]
Soh, Yeng Chai [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2008年 / 19卷 / 11期
关键词
Conic sector condition; hybrid training; real-time recurrent learning (RTRL); recurrent neural network (RNN); weight convergence;
D O I
10.1109/TNN.2008.2001923
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For a recurrent neural network (RNN), its transient response is a critical issue, especially for real-time signal processing applications. The conventional RNN training algorithms, such as backpropagation through time (BPTT) and real-time recurrent learning (RTRL), have not adequately addressed this problem because they suffer from low convergence speed. While increasing the learning rate may help to improve the performance of the RNN, it can result in unstable training in terms of weight divergence. Therefore, an optimal tradeoff between RNN training speed and weight convergence is desired. In this paper, a robust adaptive gradient-descent (RAGD) training algorithm of RNN is developed based on a novel RNN hybrid training concept. It switches the training patterns between standard real-time online backpropagation (BP) and RTRL according to the derived convergence and stability conditions. The weight convergence and L-2-stability of the algorithm are derived via the conic sector theorem. The optimized adaptive learning maximizes the training speed of the RNN for each weight update without violating the stability and convergence criteria. Computer simulations are carried out to demonstrate the applicability of the theoretical results.
引用
收藏
页码:1841 / 1853
页数:13
相关论文
共 50 条
  • [21] A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks
    Xu, Zhao
    Song, Qing
    Wang, Danwei
    NEURAL COMPUTING & APPLICATIONS, 2014, 24 (7-8): : 1851 - 1866
  • [22] A robust recurrent simultaneous perturbation stochastic approximation training algorithm for recurrent neural networks
    Zhao Xu
    Qing Song
    Danwei Wang
    Neural Computing and Applications, 2014, 24 : 1851 - 1866
  • [23] A gradient descent learning algorithm for fuzzy neural networks
    Feuring, T
    Buckley, JJ
    Hayashi, Y
    1998 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS AT THE IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE - PROCEEDINGS, VOL 1-2, 1998, : 1136 - 1141
  • [24] Training Simple Recurrent Networks through Gradient Descent Algorithms
    Castano, MA
    Casacuberta, F
    Bonet, A
    BIOLOGICAL AND ARTIFICIAL COMPUTATION: FROM NEUROSCIENCE TO TECHNOLOGY, 1997, 1240 : 493 - 500
  • [25] A data-reusing gradient descent algorithm for complex-valued recurrent neural networks
    Goh, SL
    Mandic, DP
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 2, PROCEEDINGS, 2003, 2774 : 340 - 350
  • [26] Gradient Descent Analysis: On Visualizing the Training of Deep Neural Networks
    Becker, Martin
    Lippel, Jens
    Zielke, Thomas
    PROCEEDINGS OF THE 14TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS - VOL 3: IVAPP, 2019, : 338 - 345
  • [27] A self-adaptive gradient descent search algorithm for fully-connected neural networks
    Xue, Yu
    Wang, Yankang
    Liang, Jiayu
    NEUROCOMPUTING, 2022, 478 : 70 - 80
  • [28] Explicit loss asymptotics in the gradient descent training of neural networks
    Velikanov, Maksim
    Yarotsky, Dmitry
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [29] AN ADAPTIVE CONJUGATE-GRADIENT LEARNING ALGORITHM FOR EFFICIENT TRAINING OF NEURAL NETWORKS
    ADELI, H
    HUNG, SL
    APPLIED MATHEMATICS AND COMPUTATION, 1994, 62 (01) : 81 - 102
  • [30] On Descent Spectral CG algorithms for Training Recurrent Neural Networks
    Livieris, I. E.
    Sotiropoulos, D. G.
    Pintelas, P.
    13TH PANHELLENIC CONFERENCE ON INFORMATICS, PROCEEDINGS, 2009, : 65 - +