Text Classification Based on Convolutional Neural Network and Attention Model

被引:0
作者
Yang, Shuang [1 ]
Tang, Yan [1 ]
机构
[1] Southwest Univ, Sch Comp & Informat Sci, Chongqing, Peoples R China
来源
2020 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND BIG DATA (ICAIBD 2020) | 2020年
关键词
convolutional neural network; attention model; text classification;
D O I
10.1109/icaibd49809.2020.9137447
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Using the traditional convolutional neural network (CNN) model for text classification, it is difficult to effectively capture the important local features in the text and the correlation between the feature words in the input text and the text. In order to solve this problem, this paper introduces the attention mechanism into the basic CNN model and establishes three CNN-based architectures: ATCNN-1, ATCNN-2 and ATCNN-3. ATCNN-1 increases the attention mechanism layer after the word embedding layer to obtain important local feature words, thus improving the features of convolution calculation. ATCNN-2 introduces the attention mechanism after the convolutional layer, and uses the attention mechanism to calculate the weights of each convolutional output vector to distinguish the importance degree, so that the model can extract features selectively. ATCNN-3 superimposes ATCNN-1 and ATCNN-2 together, combining the advantages of ATCNN-1 and ATCNN-2. The experimental results in the two tasks of emotion classification and Chinese news text classification show that the effects of ATCNN-1, ATCNN-2 and ATCNN-3 are obviously better than the traditional CNN model, and the classification effect is also improved to some extent compared with the classical text classification model.
引用
收藏
页码:67 / 73
页数:7
相关论文
共 23 条
[1]  
[Anonymous], 2013, 1 INT C LEARN REPR I
[2]  
Collobert R, 2011, J MACH LEARN RES, V12, P2493
[3]   Reducing the dimensionality of data with neural networks [J].
Hinton, G. E. ;
Salakhutdinov, R. R. .
SCIENCE, 2006, 313 (5786) :504-507
[4]   A fast learning algorithm for deep belief nets [J].
Hinton, Geoffrey E. ;
Osindero, Simon ;
Teh, Yee-Whye .
NEURAL COMPUTATION, 2006, 18 (07) :1527-1554
[5]   Deep Pyramid Convolutional Neural Networks for Text Categorization [J].
Johnson, Rie ;
Zhang, Tong .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, :562-570
[6]  
Joulin A., 2017, P 15 C EUR CHAPT ASS, P427, DOI DOI 10.18653/V1/E17-2068
[7]   A Convolutional Neural Network for Modelling Sentences [J].
Kalchbrenner, Nal ;
Grefenstette, Edward ;
Blunsom, Phil .
PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2014, :655-665
[8]  
Kim Y., 2014, ACL, DOI DOI 10.3115/V1/D14-1181
[9]  
Kingma Diederik P., 2015, 3 INT C LEARN REPR, DOI 10.48550/arXiv.1412.6980
[10]  
Lai SW, 2015, AAAI CONF ARTIF INTE, P2267