AELA-DLSTMs: Attention-Enabled and Location-Aware Double LSTMs for aspect-level sentiment classification

被引:37
作者
Shuang, Kai [1 ]
Ren, Xintao [1 ]
Yang, Qianqian [1 ]
Li, Rui [1 ]
Loo, Jonathan [2 ]
机构
[1] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing, Peoples R China
[2] Univ West London, Sch Comp & Engn, London, England
基金
中国国家自然科学基金;
关键词
Neural network; Long short-term memory; Attention mechanism; Aspect-level sentiment classification; MEMORY NETWORKS;
D O I
10.1016/j.neucom.2018.11.084
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Aspect-level sentiment classification, as a fine-grained task in sentiment classification, aiming to extract sentiment polarity from opinions towards a specific aspect word, has been made tremendous improvements in recent years. There are three key factors for aspect-level sentiment classification: contextual semantic information towards aspect words, correlations between aspect words and their context words, and location information of context words with regard to aspect words. In this paper, two models named AE-DLSTMs (Attention-Enabled Double LSTMs) and AELA-DLSTMs (Attention-Enabled and Location-Aware Double LSTMs) are proposed for aspect-level sentiment classification. AE-DLSTMs takes full advantage of the DLSTMs (Double LSTMs) which can capture the contextual semantic information in both forward and backward directions towards aspect words. Meanwhile, a novel attention weights generating method that combines aspect words with their contextual semantic information is designed so that those weights can make better use of the correlations between aspect words and their context words. Besides, we observe that context words with different distances or different directions towards aspect words have different contributions in sentiment polarity. Based on AE-DLSTMs, the location information of context words by assigning different weights is incorporated in AELA-DLSTMs to improve the accuracy. Experiments are conducted on two English datasets and one Chinese dataset. The experimental results have confirmed that our models can make remarkable improvements and outperform all the baseline models in all datasets, improving the accuracy of 1.67 percent to 4.77 percent in different datasets compared with baseline models. (C) 2018 Published by Elsevier B.V.
引用
收藏
页码:25 / 34
页数:10
相关论文
共 80 条
  • [1] [Anonymous], 2015, NIPS
  • [2] [Anonymous], 2016, ARXIV150906664
  • [3] [Anonymous], 2013, 2 JOINT C LEX COMP S
  • [4] [Anonymous], 2005, Kybernetes, DOI DOI 10.1108/03684920510595580
  • [5] [Anonymous], 2008, P WWW 2008 WORKSHOP
  • [6] [Anonymous], 2014, 52 ANN M ASS COMP LI
  • [7] [Anonymous], 2003, Proceedings of the eleventh ACM international conference on Multimedia, DOI [10.1145/957013.957094, DOI 10.1145/957013.957094]
  • [8] [Anonymous], 2012, ARXIV12115590
  • [9] [Anonymous], 2016, ARXIV160304351
  • [10] [Anonymous], 2014, P INT C INT C MACH L