AELA-DLSTMs: Attention-Enabled and Location-Aware Double LSTMs for aspect-level sentiment classification

被引:37
作者
Shuang, Kai [1 ]
Ren, Xintao [1 ]
Yang, Qianqian [1 ]
Li, Rui [1 ]
Loo, Jonathan [2 ]
机构
[1] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing, Peoples R China
[2] Univ West London, Sch Comp & Engn, London, England
基金
中国国家自然科学基金;
关键词
Neural network; Long short-term memory; Attention mechanism; Aspect-level sentiment classification; MEMORY NETWORKS;
D O I
10.1016/j.neucom.2018.11.084
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Aspect-level sentiment classification, as a fine-grained task in sentiment classification, aiming to extract sentiment polarity from opinions towards a specific aspect word, has been made tremendous improvements in recent years. There are three key factors for aspect-level sentiment classification: contextual semantic information towards aspect words, correlations between aspect words and their context words, and location information of context words with regard to aspect words. In this paper, two models named AE-DLSTMs (Attention-Enabled Double LSTMs) and AELA-DLSTMs (Attention-Enabled and Location-Aware Double LSTMs) are proposed for aspect-level sentiment classification. AE-DLSTMs takes full advantage of the DLSTMs (Double LSTMs) which can capture the contextual semantic information in both forward and backward directions towards aspect words. Meanwhile, a novel attention weights generating method that combines aspect words with their contextual semantic information is designed so that those weights can make better use of the correlations between aspect words and their context words. Besides, we observe that context words with different distances or different directions towards aspect words have different contributions in sentiment polarity. Based on AE-DLSTMs, the location information of context words by assigning different weights is incorporated in AELA-DLSTMs to improve the accuracy. Experiments are conducted on two English datasets and one Chinese dataset. The experimental results have confirmed that our models can make remarkable improvements and outperform all the baseline models in all datasets, improving the accuracy of 1.67 percent to 4.77 percent in different datasets compared with baseline models. (C) 2018 Published by Elsevier B.V.
引用
收藏
页码:25 / 34
页数:10
相关论文
共 80 条
[51]   Deep Sentence Embedding Using Long Short-Term Memory Networks: Analysis and Application to Information Retrieval [J].
Palangi, Hamid ;
Deng, Li ;
Shen, Yelong ;
Gao, Jianfeng ;
He, Xiaodong ;
Chen, Jianshu ;
Song, Xinying ;
Ward, Rabab .
IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2016, 24 (04) :694-707
[52]   Thumbs up? Sentiment classification using machine learning techniques [J].
Pang, B ;
Lee, L ;
Vaithyanathan, S .
PROCEEDINGS OF THE 2002 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, 2002, :79-86
[53]  
Pennington Jeffrey, 2014, P 2014 C EMP METH NA, P1532
[54]  
Pontiki M., 2014, P 8 INT WORKSHOP SEM, P27
[55]  
Pontiki M., 2016, P 10 INT WORKSH SEM, P19, DOI [DOI 10.18653/V1/S16-1002, 10.18653/v1/S16-1002]
[56]  
Ren YF, 2016, AAAI CONF ARTIF INTE, P215
[57]  
Riedel S., 2017, P 2017 C EMPIRICAL M, P452, DOI DOI 10.18653/V1/D17-1047
[58]  
Rosenthal S., 2017, P 11 INT WORKSHOP SE, P502, DOI [10.18653/v1/S17-2088, DOI 10.18653/V1/S17-2088]
[59]  
Sak H., 2014, P 15 ANN C INT SPEEC
[60]  
Socher R., 2013, PROC 26 INT C NEURAL, P926