Fast Training of Deep LSTM Networks

被引:7
作者
Yu, Wen [1 ]
Li, Xiaoou [2 ]
Gonzalez, Jesus [1 ]
机构
[1] CINVESTAV IPN Natl Polytech Inst, Dept Control Automat, Mexico City, DF, Mexico
[2] CINVESTAV IPN Natl Polytech Inst, Dept Comp, Mexico City, DF, Mexico
来源
ADVANCES IN NEURAL NETWORKS - ISNN 2019, PT I | 2019年 / 11554卷
关键词
NEURAL-NETWORKS; SYSTEMS;
D O I
10.1007/978-3-030-22796-8_1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep recurrent neural networks (RNN), such as LSTM, have many advantages over forward networks. However, the LSTM training method, such as backward propagation through time (BPTT), is really slow. In this paper, by separating the LSTM cell into forward and recurrent substructures, we propose a much simpler and faster training method than the BPTT. The deep LSTM is modified by combining the deep RNN with the multilayer perceptron (MLP). The simulation results show that our fast training method for LSTM is better than BPTT for LSTM.
引用
收藏
页码:3 / 10
页数:8
相关论文
共 26 条
[1]  
[Anonymous], 2013, NONLINEAR SYSTEM IDE
[2]  
[Anonymous], 2012, ADV NEURAL INFORM PR
[3]  
Bengio P., 2006, Advances in Neural Information Processing Systems 19 (NIPS06), P153, DOI DOI 10.5555/2976456.2976476
[4]  
Box G.E.P., 2008, TIME SERIES ANAL
[5]   A review on neural networks with random weights [J].
Cao, Weipeng ;
Wang, Xizhao ;
Ming, Zhong ;
Gao, Jinzhu .
NEUROCOMPUTING, 2018, 275 :278-287
[6]  
Chung J, 2014, Em-
[7]  
Graves A, 2012, STUD COMPUT INTELL, V385, P1, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
[8]  
Graves A, 2013, INT CONF ACOUST SPEE, P6645, DOI 10.1109/ICASSP.2013.6638947
[9]   A fast learning algorithm for deep belief nets [J].
Hinton, Geoffrey E. ;
Osindero, Simon ;
Teh, Yee-Whye .
NEURAL COMPUTATION, 2006, 18 (07) :1527-1554
[10]  
Hirose Noriaki, 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA), P6471, DOI 10.1109/ICRA.2017.7989764