A RNN based Approach for next word prediction in Assamese Phonetic Transcription

被引:19
作者
Barman, Partha Pratim [1 ]
Boruah, Abhijit [1 ]
机构
[1] Dibrugarh Univ, DUIET, Dept CSE, Dibrugarh 786004, Assam, India
来源
8TH INTERNATIONAL CONFERENCE ON ADVANCES IN COMPUTING & COMMUNICATIONS (ICACC-2018) | 2018年 / 143卷
关键词
Next word prediction; LSTM; RNN; Assamese;
D O I
10.1016/j.procs.2018.10.359
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In this paper, we present a Long Short Term Memory network (LSTM) model which is a special kind of Recurrent Neural Network(RNN) for instant messaging, where the goal is to predict next word(s) given a set of current words to the user. This method is more complex in other languages apart from English. For instance, in Assamese language, there are some equivalent synonyms of the word 'you', that is used to address a second person in English. Here, we have developed a solution to this issue by storing the transcripted Assamese language according to International Phonetic Association(IPA) chart and have fed the data into our model. Our model goes through the data set of the transcripted Assamese words and predicts the next word using LSTM with an accuracy of 88.20% for Assamese text and 72.10% for phonetically transcripted Assamese language. This model can be used in predicting next word of Assamese language, especially at the time of phonetic typing. (C) 2018 The Authors. Published by Elsevier B.V.
引用
收藏
页码:117 / 123
页数:7
相关论文
共 9 条
[1]  
[Anonymous], 2014, Generating sequences with recurrent neural networks
[2]  
[Anonymous], 2015, ADV NEURAL INFORM PR
[3]  
Bahdanau D., 2015, Neural machine translation
[4]   LEARNING LONG-TERM DEPENDENCIES WITH GRADIENT DESCENT IS DIFFICULT [J].
BENGIO, Y ;
SIMARD, P ;
FRASCONI, P .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (02) :157-166
[5]  
Gers F. A., 1999, LEARNING FORGET CONT
[6]  
Hochreiter S, 1997, ADV NEUR IN, V9, P473
[7]  
Mikolov T, 2010, 11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 1-2, P1045
[8]  
Olah C., 2015, colah. github. io
[9]  
Zhou C., 2015, COMPUT ENCE