Dependency-Tree Based Convolutional Neural Networks for Aspect Term Extraction

被引:34
作者
Ye, Hai [1 ]
Yan, Zichao [1 ]
Luo, Zhunchen [2 ]
Chao, Wenhan [1 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, Beijing, Peoples R China
[2] China Def Sci & Technol Informat Ctr, Beijing, Peoples R China
来源
ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2017, PT II | 2017年 / 10235卷
基金
中国国家自然科学基金;
关键词
Aspect term extraction; Dependency information; Tree-based convolution; Deep learning;
D O I
10.1007/978-3-319-57529-2_28
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Aspect term extraction is one of the fundamental subtasks in aspect-based sentiment analysis. Previous work has shown that sentences' dependency information is critical and has been widely used for opinion mining. With recent success of deep learning in natural language processing (NLP), recurrent neural network (RNN) has been proposed for aspect term extraction and shows the superiority over feature-rich CRFs based models. However, because RNN is a sequential model, it can not effectively capture tree-based dependency information of sentences thus limiting its practicability. In order to effectively exploit sentences' dependency information and leverage the effectiveness of deep learning, we propose a novel dependency-tree based convolutional stacked neural network (DTBCSNN) for aspect term extraction, in which tree-based convolution is introduced over sentences' dependency parse trees to capture syntactic features. Our model is an end-to-end deep learning based model and it does not need any human-crafted features. Furthermore, our model is flexible to incorporate extra linguistic features to further boost the model performance. To substantiate, results from experiments on SemEval2014 Task4 datasets (reviews on restaurant and laptop domain) show that our model achieves outstanding performance and outperforms the RNN and CRF baselines.
引用
收藏
页码:350 / 362
页数:13
相关论文
共 23 条
  • [1] [Anonymous], 2005, Proceedings of the ACM international conference on world wide web, DOI 10.1145/1060745.1060797
  • [2] [Anonymous], 2010, P 14 C COMPUTATIONAL
  • [3] Chernyshevich M., 2014, Proceedings of SemEval, P309
  • [4] Collobert R, 2011, J MACH LEARN RES, V12, P2493
  • [5] Hu M, 2004, P 10 ACM SIGKDD INT, P168, DOI DOI 10.1145/1014052.1014073
  • [6] Irsoy O, 2014, EMNLP, P720, DOI DOI 10.3115/V1/D14-1080
  • [7] Jakob N., 2010, Proceedings of the 2010 conference on empirical methods in natural language processing, P1035
  • [8] Jin W., 2009, ICME INT C COMPL MED, P1, DOI [10.1145/1553374.1553435, DOI 10.1145/1553374.1553435, DOI 10.1109/ICCME.2009.4906624, 10.1109/ICCME.2009.4906624]
  • [9] Johansson Richard., 2011, P 49 ANN M ASS COMPU, P101
  • [10] Kim Y., 2014, ARXIV 14085882, DOI DOI 10.3115/V1/D14-1181