Robust Adaptive Gradient-Descent Training Algorithm for Recurrent Neural Networks in Discrete Time Domain

被引:68
|
作者
Song, Qing [1 ]
Wu, Yilei [1 ]
Soh, Yeng Chai [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2008年 / 19卷 / 11期
关键词
Conic sector condition; hybrid training; real-time recurrent learning (RTRL); recurrent neural network (RNN); weight convergence;
D O I
10.1109/TNN.2008.2001923
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For a recurrent neural network (RNN), its transient response is a critical issue, especially for real-time signal processing applications. The conventional RNN training algorithms, such as backpropagation through time (BPTT) and real-time recurrent learning (RTRL), have not adequately addressed this problem because they suffer from low convergence speed. While increasing the learning rate may help to improve the performance of the RNN, it can result in unstable training in terms of weight divergence. Therefore, an optimal tradeoff between RNN training speed and weight convergence is desired. In this paper, a robust adaptive gradient-descent (RAGD) training algorithm of RNN is developed based on a novel RNN hybrid training concept. It switches the training patterns between standard real-time online backpropagation (BP) and RTRL according to the derived convergence and stability conditions. The weight convergence and L-2-stability of the algorithm are derived via the conic sector theorem. The optimized adaptive learning maximizes the training speed of the RNN for each weight update without violating the stability and convergence criteria. Computer simulations are carried out to demonstrate the applicability of the theoretical results.
引用
收藏
页码:1841 / 1853
页数:13
相关论文
共 50 条
  • [32] Adaptive learning with guaranteed stability for discrete-time recurrent neural networks
    邓华
    吴义虎
    段吉安
    JournalofCentralSouthUniversityofTechnology, 2007, (05) : 685 - 689
  • [33] Adaptive learning with guaranteed stability for discrete-time recurrent neural networks
    Hua Deng
    Yi-hu Wu
    Ji-an Duan
    Journal of Central South University of Technology, 2007, 14 : 685 - 689
  • [34] Adaptive learning with guaranteed stability for discrete-time recurrent neural networks
    Deng Hua
    Wu Yi-hu
    Duan Ji-an
    JOURNAL OF CENTRAL SOUTH UNIVERSITY OF TECHNOLOGY, 2007, 14 (05): : 685 - 689
  • [35] Fast gradient descent algorithm for image classification with neural networks
    Abdelkrim El Mouatasim
    Signal, Image and Video Processing, 2020, 14 : 1565 - 1572
  • [36] Training qubit neural network with hybrid genetic algorithm and gradient descent for indirect adaptive controller design
    Ganjefar, Soheil
    Tofighi, Morteza
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2017, 65 : 346 - 360
  • [37] A Training Algorithm for Locally Recurrent Neural Networks Based on the Explicit Gradient of the Loss Function
    Carcangiu, Sara
    Montisci, Augusto
    ALGORITHMS, 2025, 18 (02)
  • [38] A training algorithm for partial recurrent neural networks
    Al-Faysale, MSM
    Modelling and Simulation 2004, 2004, : 350 - 354
  • [39] Robust passivity analysis for discrete-time recurrent neural networks with mixed delays
    Huang, Chuan-Kuei
    Shu, Yu-Jeng
    Chang, Koan-Yuh
    Shou, Ho-Nien
    Lu, Chien-Yu
    INTERNATIONAL JOURNAL OF ELECTRONICS, 2015, 102 (02) : 216 - 232
  • [40] A robust evolutionary algorithm for training neural networks
    Yang, JM
    Kao, CY
    NEURAL COMPUTING & APPLICATIONS, 2001, 10 (03): : 214 - 230