Injecting User Identity Into Pretrained Language Models for Document-Level Sentiment Classification

被引:5
作者
Cao, Xinlei [1 ]
Yu, Jinyang [2 ]
Zhuang, Yan [3 ,4 ]
机构
[1] East China Normal Univ, Sch Comp Sci & Technol, Shanghai 200062, Peoples R China
[2] Ctr ADR Monitoring Guangdong, Guangzhou 510080, Peoples R China
[3] Chinese Peoples Liberat Army Gen Hosp, Natl Engn Lab Med Big Data Applicat Technol, Beijing 100853, Peoples R China
[4] Chinese Peoples Liberat Army Gen Hosp, Med Big Data Res Ctr, Med Innovat Res Div, Beijing 100853, Peoples R China
关键词
Task analysis; Bit error rate; Predictive models; Transformers; Context modeling; Electronic mail; Data models; Representation learning; document-level sentiment classification; personalized sentiment classification; pre-trained language models; attention mechanism; EMBEDDINGS;
D O I
10.1109/ACCESS.2022.3158975
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper mainly studies the combination of pre-trained language models and user identity information for document-level sentiment classification. In recent years, pre-trained language models (PLMs) such as BERT have achieved state-of-the-art results on many NLP applications, including document-level sentiment classification. On the other hand, a collection of works introduce additional information such as user identity for better text modeling. However, most of them inject user identity into traditional models, while few studies have been conducted to study the combination of pre-trained language models and user identity for even better performance. To address this issue, in this paper, we propose to unite user identity and PLMs and formulate User-enhanced Pre-trained Language Models (U-PLMs). Specifically, we demonstrate two simple yet effective attempts, i.e. embedding-based and attention-based personalization, which inject user identity into different parts of a pre-trained language model and provide personalization from different perspectives. Experiments in three datasets with two backbone PLMs show that our proposed methods outperform the best state-of-the-art baseline method with an absolute improvement of up to 3%, 2.8%, and 2.2% on accuracy. In addition, our methods encode user identity with plugin modules, which are fully compatible with most auto-encoding pre-trained language models.
引用
收藏
页码:30157 / 30167
页数:11
相关论文
共 40 条
[1]  
Amplayo RK, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, P2535
[2]  
Amplayo RK, 2019, P 2019 C EMP METH NA, P5602, DOI [10.18653/v1/D19-1562, DOI 10.18653/V1/D19-1562]
[3]  
Ba JimmyLei., 2016, CORR
[4]  
Bhatta J, 2020, Journal of Innovations in Engineering Education, V3, P71, DOI [10.3126/jiee.v3i1.34327, 10.3126/jiee.v3i1.34327, DOI 10.3126/JIEE.V3I1.34327]
[5]  
Brown TB, 2020, ADV NEUR IN, V33
[6]  
Chen Huimin., 2016, P 2016 C EMPIRICAL M, P1650, DOI [10.18653/v1/D16-1171, DOI 10.18653/V1/D16-1171]
[7]   AUDIO ALBERT: A LITE BERT FOR SELF-SUPERVISED LEARNING OF AUDIO REPRESENTATION [J].
Chi, Po-Han ;
Chung, Pei-Hung ;
Wu, Tsung-Han ;
Hsieh, Chun-Cheng ;
Chen, Yen-Hao ;
Li, Shang-Wen ;
Lee, Hung-yi .
2021 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP (SLT), 2021, :344-350
[8]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[9]  
Finn C, 2017, PR MACH LEARN RES, V70
[10]  
Fu P, 2018, AAAI CONF ARTIF INTE, P4808