A Hybrid Bidirectional Recurrent Convolutional Neural Network Attention-Based Model for Text Classification

被引:41
作者
Zheng, Jin [2 ]
Zheng, Limin [1 ,2 ]
机构
[1] China Agr Univ, Coll Informat & Elect Engn, Beijing 100083, Peoples R China
[2] China Agr Univ, Beijing Lab Food Qual & Safety, Beijing 100083, Peoples R China
关键词
Attention mechanism; bidirectional long short-term memory; convolutional neural network; fine-grained sentiment analysis; multi-class text classification;
D O I
10.1109/ACCESS.2019.2932619
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The text classification task is an important application in natural language processing. At present, deep learning models, such as convolutional neural network and recurrent neural network, have achieved good results for this task, but the multi-class text classification and the fine-grained sentiment analysis are still challenging. In this paper, we propose a hybrid bidirectional recurrent convolutional neural network attention-based model to address this issue, which named BRCAN. The model combines the bidirectional long short-term memory and the convolutional neural network with the attention mechanism and word2vec to achieve the fine-grained text classification task. In our model, we apply word2vec to generate word vectors automatically and a bidirectional recurrent structure to capture contextual information and long-term dependence of sentences. We also employ a maximum pool layer of convolutional neural network that judges which words play an essential role in text classification, and use the attention mechanism to give them higher weights to capture the key components in texts. We conduct experiments on four datasets, including Yahoo! Answers, Sogou News of the topic classification, Yelp Reviews, and Douban Movies Top250 short reviews of the sentiment analysis. And the experimental results show that the BRCAN outperforms the state-of-the-art models.
引用
收藏
页码:106673 / 106685
页数:13
相关论文
共 40 条
[1]  
[Anonymous], 2010, Proceedings of the 23rd International Conference on Computational Linguistics
[2]  
[Anonymous], 2012, Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Short Papers - Volume 2, ACL '12
[3]  
[Anonymous], P EMNLP
[4]  
Bridle J. S., 1990, Neurocomputing, P227
[5]  
Conneau A, 2017, 15TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2017), VOL 1: LONG PAPERS, P1107
[6]   Attention pooling-based convolutional neural network for sentence modelling [J].
Er, Meng Joo ;
Zhang, Yong ;
Wang, Ning ;
Pratama, Mahardhika .
INFORMATION SCIENCES, 2016, 373 :388-403
[7]  
Fan RE, 2008, J MACH LEARN RES, V9, P1871
[8]  
Gao S, 2018, REPRESENTATION LEARNING FOR NLP, P11
[9]  
Gers FA, 1999, IEE CONF PUBL, P850, DOI [10.1162/089976600300015015, 10.1049/cp:19991218]
[10]  
Graves A, 2013, INT CONF ACOUST SPEE, P6645, DOI 10.1109/ICASSP.2013.6638947