共 28 条
Nonmonotone Levenberg-Marquardt training of recurrent neural architectures for processing symbolic sequences
被引:7
|作者:
Peng, Chun-Cheng
[1
]
Magoulas, George D.
[1
]
机构:
[1] Univ London, Birkbeck Coll, Dept Comp Sci & Informat Syst, London WC1E 7HX, England
关键词:
Levenberg-Marquardt methods;
Nonmonotone learning;
Recurrent neural networks;
ALGORITHM;
CONVERGENCE;
NETWORKS;
D O I:
10.1007/s00521-010-0493-2
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
In this paper, we present nonmonotone variants of the Levenberg-Marquardt (LM) method for training recurrent neural networks (RNNs). These methods inherit the benefits of previously developed LM with momentum algorithms and are equipped with nonmonotone criteria, allowing temporal increase in training errors, and an adaptive scheme for tuning the size of the nonmonotone slide window. The proposed algorithms are applied to training RNNs of various sizes and architectures in symbolic sequence-processing problems. Experiments show that the proposed nonmonotone learning algorithms train more effectively RNNs for sequence processing than the original monotone methods.
引用
收藏
页码:897 / 908
页数:12
相关论文