The New Large-Scale RNNLM System Based On Distributed Neuron

被引:3
作者
Niu, Dejiao [1 ]
Xue, Rui [1 ]
Cai, Tao [1 ]
Li, Hai [2 ]
Effah, Kingsley [1 ]
Zhang, Hang [1 ]
机构
[1] Jiangsu Univ, Sch Comp Sci & Commun Engn, Zhenjiang, Jiangsu, Peoples R China
[2] Duke Univ, Elect & Comp Engn, Durham, NC 27706 USA
来源
2017 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW) | 2017年
基金
中国博士后科学基金;
关键词
Recurrent Neural Network Language Model; Distributed Computing; Distributed System; Spark;
D O I
10.1109/IPDPSW.2017.21
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
RNNLM (Recurrent Neural Network Language Model) can save the historical information of the training dataset by the last hidden layer and can also as input for training. It has become an interesting topic in the field of Natural Language Processing research. However, the immense training time overhead is a big problem. The large output layer, hidden layer, last hidden layer and the connections among them will generate enormous matrix in training. It is the main facts to influence the efficiency and scalability. At the same time, output layer class and small hidden layer should decrease the accuracy of RNNLM. In general, the lack of parallel for artificial neuron is main reason for these. We change the structure of RNNLM and design the new large-scale RNNLM by the center of distributed artificial neurons in hidden layer to stimulate the parallel characteristic of biological neuron system. Meanwhile, we change training method, and present the coordination strategy for distributed neuron. At last, the prototype of new large-scale RNNLM system is implemented based on Spark. The testing and analysis results show that the training time overhead is far less than the growth rate of the distributed neuron in hidden layer and size of training dataset. These results show our large-scale RNNLM system has efficiency and scalability advantage.
引用
收藏
页码:433 / 436
页数:4
相关论文
共 11 条
[1]  
[Anonymous], 2011, IEEE INT C AC SPEECH
[2]  
[Anonymous], 2012, INT C MACH LEARN
[3]   A neural probabilistic language model [J].
Bengio, Y ;
Ducharme, R ;
Vincent, P ;
Jauvin, C .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (06) :1137-1155
[4]  
He B., SOCC 10
[5]  
Iandola ForrestN., 2013, ICIP
[6]  
Li Boxun, LARGE SCALE RECURREN
[7]  
Mikolov T, 2010, 11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 1-2, P1045
[8]  
Ng Andrew Y., DEEP LEARNING COTS H
[9]  
Pauls D., 2011, P ACL HLT, P258
[10]  
Tomas Mikolov, 2012, THESIS