Attribute-Based Injection Transformer for Personalized Sentiment Analysis

被引:2
作者
Zhang, You [1 ]
Wang, Jin [1 ]
Yu, Liang-Chih [2 ]
Xu, Dan [1 ]
Zhang, Xuejie [1 ]
机构
[1] Yunnan Univ, Sch Informat Sci & Engn, Kunming 650000, Peoples R China
[2] Yuan Ze Univ, Dept Informat Management, Taoyuan 320, Taiwan
来源
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE | 2024年 / 8卷 / 03期
基金
中国国家自然科学基金;
关键词
Transformers; Reviews; Sentiment analysis; Task analysis; Analytical models; Context modeling; Training; Personalized sentiment analysis; attention mechanism; layer normalization; pre-trained language model; CLASSIFICATION;
D O I
10.1109/TETCI.2024.3369323
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Personal attributes have been proven to be useful for sentiment analysis. However, previous models of learning attribute-specific language representations are suboptimal because only context- or content-wise injection is adopted. This study proposes a transformer structure with a combination of both context- and content-wise injections based on a well pretrained transformer encoder. For context-wise injection, self-interactive attention is implemented by incorporating personal attributes into a multi-head attention. For the content-wise perspective, an attribute-based layer normalization is used to align text representation with personal attributes. In particular, the proposed transformer layer can be a universal layer compatible with the original Google Transformer layer. Instead of training from scratch, the proposed Transformer layer can be initialized from a well pre-trained checkpoint for downstream tasks. Extensive experiments were conducted on three benchmarks of document-level sentiment analysis, including IMDB, Yelp-2013 and Yelp-2014. The results show that the proposed method outperforms the previous methods for personalized sentiment analysis, demonstrating that the combination of both context- and content-wise injections can facilitate model learning for attribute-specific language representations.
引用
收藏
页码:2581 / 2591
页数:11
相关论文
共 38 条
  • [1] Amplayo RK, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P5602
  • [2] Ba Jimmy Lei, 2016, Layer normalization
  • [3] Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
  • [4] An Attentive Survey of Attention Models
    Chaudhari, Sneha
    Mithal, Varun
    Polatkan, Gungor
    Ramanath, Rohan
    [J]. ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2021, 12 (05)
  • [5] Chen H, 2016, P 2016 C EMP METH NA, P1650, DOI DOI 10.18653/V1/D16-1171
  • [6] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [7] Dong L, 2017, 15TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2017), VOL 1: LONG PAPERS, P623
  • [8] Graves A, 2012, STUD COMPUT INTELL, V385, P1, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
  • [9] Pre-trained models: Past, present and future
    Han, Xu
    Zhang, Zhengyan
    Ding, Ning
    Gu, Yuxian
    Liu, Xiao
    Huo, Yuqi
    Qiu, Jiezhong
    Yao, Yuan
    Zhang, Ao
    Zhang, Liang
    Han, Wentao
    Huang, Minlie
    Jin, Qin
    Lan, Yanyan
    Liu, Yang
    Liu, Zhiyuan
    Lu, Zhiwu
    Qiu, Xipeng
    Song, Ruihua
    Tang, Jie
    Wen, Ji-Rong
    Yuan, Jinhui
    Zhao, Wayne Xin
    Zhu, Jun
    [J]. AI OPEN, 2021, 2 : 225 - 250
  • [10] Ioffe Sergey, 2015, INT C MACH LEARN, V37, P448, DOI DOI 10.48550/ARXIV.1502.03167