A recalling-enhanced recurrent neural network: Conjugate gradient learning algorithm and its convergence analysis

被引:57
作者
Gao, Tao [1 ,2 ]
Gong, Xiaoling [1 ,2 ]
Zhang, Kai [3 ]
Lin, Feng [2 ,4 ]
Wang, Jian [2 ]
Huang, Tingwen [5 ]
Zurada, Jacek M. [6 ,7 ]
机构
[1] China Univ Petr East China, Coll Control Sci & Engn, Qingdao 266580, Peoples R China
[2] China Univ Petr East China, Coll Sci, Qingdao 266580, Peoples R China
[3] China Univ Petr East China, Sch Petr Engn, Qingdao 266580, Peoples R China
[4] Beijing Sunwad Sci & Technol Ltd, Beijing 100124, Peoples R China
[5] Texas A&M Univ Qatar, Doha 23874, Qatar
[6] Univ Louisville, Elect & Comp Engn, Louisville, KY 40292 USA
[7] Univ Social Sci, Informat Technol Inst, PL-90113 Lodz, Poland
基金
中国国家自然科学基金;
关键词
Recurrent; Neural network; Conjugate gradient; Generalized Armijo search; Monotonicity; Convergence; DETERMINISTIC CONVERGENCE; MINIMIZATION;
D O I
10.1016/j.ins.2020.01.045
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Elman network is a classical recurrent neural network with an internal delay feedback. In this paper, we propose a recalling-enhanced recurrent neural network (RERNN) which has a selective memory property. In addition, an improved conjugate algorithm with generalized Armijo search technique that speeds up the convergence rate is used to train the RERNN model. Further enhancement performance is achieved with adaptive learning coefficients. Finally, we prove weak and strong convergence of the presented algorithm. In other words, as the number of training steps increases, the following has been established for RERNN: (1) the gradient norm of the error function with respect to the weight vectors converges to zero, (2) the weight sequence approaches a fixed optimal point. We have carried out a number of simulations to illustrate and verify the theoretical results that demonstrate the efficiency of the proposed algorithm. (C) 2020 Elsevier Inc. All rights reserved.
引用
收藏
页码:273 / 288
页数:16
相关论文
共 43 条
[1]  
[Anonymous], 2017, EXPT STUDY ELMAN NET
[2]  
[Anonymous], 1952, METHOD CONJUGATE GRA
[3]  
[Anonymous], INT C BIOINSP COMP T
[5]   A spectral conjugate gradient method for unconstrained optimization [J].
Birgin, EG ;
Martínez, JM .
APPLIED MATHEMATICS AND OPTIMIZATION, 2001, 43 (02) :117-128
[6]   On the implicit acquisition of a context-free grammar by a simple recurrent neural network [J].
Cartling, Bo .
NEUROCOMPUTING, 2008, 71 (7-9) :1527-1537
[7]   A nonlinear conjugate gradient method with a strong global convergence property [J].
Dai, YH ;
Yuan, Y .
SIAM JOURNAL ON OPTIMIZATION, 1999, 10 (01) :177-182
[8]   A transmission line fault locator based on Elman recurrent networks [J].
Ekici, Sami ;
Yildirim, Selcuk ;
Poyraz, Mustafa .
APPLIED SOFT COMPUTING, 2009, 9 (01) :341-347
[9]   FINDING STRUCTURE IN TIME [J].
ELMAN, JL .
COGNITIVE SCIENCE, 1990, 14 (02) :179-211
[10]   FUNCTION MINIMIZATION BY CONJUGATE GRADIENTS [J].
FLETCHER, R ;
REEVES, CM .
COMPUTER JOURNAL, 1964, 7 (02) :149-&