Recurrent neural network with attention mechanism for language model

被引:0
作者
Mu-Yen Chen
Hsiu-Sen Chiang
Arun Kumar Sangaiah
Tsung-Che Hsieh
机构
[1] National Taichung University of Science and Technology,Department of Information Management
[2] Vellore Institute of Technology,School of Computer Science and Engineering
来源
Neural Computing and Applications | 2020年 / 32卷
关键词
Language model; Recurrent neural network; Artificial intelligence; Attention mechanism;
D O I
暂无
中图分类号
学科分类号
摘要
The rapid growth of the Internet promotes the growth of textual data, and people get the information they need from the amount of textual data to solve problems. The textual data may include some potential information like the opinions of the crowd, the opinions of the product, or some market-relevant information. However, some problems that point to “How to get features from the text” must be solved. The model of extracting the text features by using the neural network method is called neural network language model. The features are based on n-gram Model concept, which are the co-occurrence relationship between the vocabularies. The word vectors are important because the sentence vectors or the document vectors still have to understand the relationship between the words, and based on this, this study discusses the word vectors. This study assumes that the words contain “the meaning in sentences” and “the position of grammar.” This study uses recurrent neural network with attention mechanism to establish a language model. This study uses Penn Treebank, WikiText-2, and NLPCC2017 text datasets. According to these datasets, the proposed models provide the better performance by the perplexity.
引用
收藏
页码:7915 / 7923
页数:8
相关论文
共 24 条
[1]  
Khodabakhsh M(2018)Predicting future personal life events on twitter via recurrent neural networks J Intell Inf Syst 49 44-52
[2]  
Kahani M(2018)Augmented sentiment representation by learning context information Neural Comput Appl 3 1137-1155
[3]  
Bagheri E(2019)Hierarchical attention based long short-term memory for Chinese lyric generation Appl Intell 30 261-270
[4]  
Han H(2003)A neural probabilistic language model J Mach Learn Res 30 607-616
[5]  
Bai X(2018)Mine blast algorithm for environmental economic load dispatch with valve loading effect Neural Comput Appl 28 353-364
[6]  
Li P(2018)Load frequency controller design of a two-area system composing of PV grid and thermal generator via firefly algorithm Neural Comput Appl 19 313-330
[7]  
Wu X(2017)PI controller design using ABC algorithm for MPPT of PV system supplying DC motor-pump load Neural Comput Appl undefined undefined-undefined
[8]  
Du Z(1993)Building a large annotated corpus of English: the Penn Treebank Computational linguistics undefined undefined-undefined
[9]  
Guo Y(undefined)undefined undefined undefined undefined-undefined
[10]  
Fujita H(undefined)undefined undefined undefined undefined-undefined