Sentiment aware word emb e ddings using refinement and senti-contextualized learning approach

被引:26
作者
Naderalvojoud, Behzad [1 ]
Sezer, Ebru Akcapinar [1 ]
机构
[1] Hacettepe Univ, Comp Engn Dept, TR-06800 Beytepe, Turkey
关键词
Word embedding; Sentiment analysis; Deep learning; EMBEDDINGS;
D O I
10.1016/j.neucom.2020.03.094
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most pre-trained word embeddings are achieved from context-based learning algorithms trained over a large text corpus. This leads to learning similar vectors for words that share most of their contexts, while expressing different meanings. Therefore, the complex characteristics of words cannot be fully learned by using such models. One of the natural language processing applications that suffers from this problem is sentiment analysis. In this task, two words with opposite sentiments are not distinguished well by using common pre-trained word embeddings. This paper addresses this problem and proposes two simple, but empirically effective, approaches to learn word embeddings for sentiment analysis. The both approaches exploit sentiment lexicons and take into account the polarity of words in learning word embeddings. While the first approach encodes the sentiment information of words into existing pre-trained word embeddings, the second one builds synthetic sentimental contexts for embedding models along with other semantic contexts. The word embeddings obtained from the both approaches are evaluated on several sentiment classification tasks using Skip-gram and GloVe models. Results show that both approaches improve state-of-the-art results using basic deep learning models over sentiment analysis benchmarks. (c) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页码:149 / 160
页数:12
相关论文
共 50 条
[1]   Deep learning-based sentiment classification of evaluative text based on Multi-feature fusion [J].
Abdi, Asad ;
Shamsuddin, Siti Mariyam ;
Hasan, Shafaatunnur ;
Piran, Jalil .
INFORMATION PROCESSING & MANAGEMENT, 2019, 56 (04) :1245-1259
[2]  
[Anonymous], 2015, ARXIV150505008
[3]  
[Anonymous], 2014, Transactions of the Association for Computational Linguistics, DOI DOI 10.1162/TACLA00162
[4]  
[Anonymous], ARXIV180704990
[5]  
[Anonymous], 2013, P 2013 C EMPIRICAL M
[6]  
[Anonymous], 2017, P 31 AAAI C ART INT
[7]  
[Anonymous], 2013, 2 JOINT C LEX COMP S
[8]  
[Anonymous], 2014, COLING
[9]  
Baccianella S., 2010, LREC 2010 7 INT C LA, P2200
[10]  
Bengio Y, 2001, ADV NEUR IN, V13, P932