Adaptive Pre-Training and Collaborative Fine-Tuning: A Win-Win Strategy to Improve Review Analysis Tasks

被引:1
作者
Mao, Qianren [1 ,2 ]
Li, Jianxin [1 ,2 ]
Lin, Chenghua [3 ]
Chen, Congwen [4 ]
Peng, Hao [1 ,2 ]
Wang, Lihong [1 ,2 ]
Yu, Philip S. [5 ]
机构
[1] Beihang Univ, Beijing Adv Innovat Ctr Big Data & Brain Comp, Beijing 100083, Peoples R China
[2] Beihang Univ, State Key Lab Software Dev Environm, Beijing 100083, Peoples R China
[3] Univ Sheffield, Dept Comp Sci, Sheffield S10 2TN, S Yorkshire, England
[4] Delft Univ Technol, Fac EEMCS, NL-2628 CD Delft, Netherlands
[5] Univ Illinois, Dept Comp Sci, Chicago, IL 60607 USA
基金
英国工程与自然科学研究理事会;
关键词
Task analysis; Multitasking; Adaptation models; Collaboration; Training; Predictive models; Context modeling; Pre-training; review analysis; review summarization; RoBERTa; sentiment classification; task-adaptive;
D O I
10.1109/TASLP.2022.3140482
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Summarizing user reviews and classifying user sentiment are two critical tasks for modern e-commerce platforms. These two tasks can benefit each other by capturing the shared linguistic features. However, such a relationship has not been fully exploited by existing research on domain-specific contextual representations. This work explores a win-win strategy for a multi-task framework with three stages: general pre-training, adaptive pre-training, and collaborative fine-tuning. The task-adaptive continual pre-training on a language model can obtain domain-specific contextual representations, further used to improve two related tasks, sentiment classification and review summarization during the collaborative fine-tuning. Meanwhile, to effectively capture sentiment-oriented domain-specific contextual representations, we introduce a novel task-adaptive pre-training procedure, which adds a sentiment prediction task during the adaptive pre-training. Extensive experiments conducted on two adaption scenarios of a general-to-single domain and a general-to-multiple domain show that our framework outperforms state-of-the-art methods.
引用
收藏
页码:622 / 634
页数:13
相关论文
共 34 条
[1]  
[Anonymous], 2010, P 23 INT C COMP LING
[2]  
[Anonymous], 2004, P 10 ACM SIGKDD INT
[3]  
[Anonymous], 2014, P COLING 2014 25 INT
[4]  
Brodersen Kay H., 2010, Proceedings of the 2010 20th International Conference on Pattern Recognition (ICPR 2010), P3121, DOI 10.1109/ICPR.2010.764
[5]   A Unified Dual-view Model for Review Summarization and Sentiment Classification with Inconsistency Loss [J].
Chan, Hou Pong ;
Chen, Wang ;
King, Irwin .
PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, :1191-1200
[6]  
de Marneffe MC, 2014, LREC 2014 - NINTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, P4585
[7]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[8]  
Gu Y., 2020, ARXIV200409733
[9]  
Gururangan Suchin, 2020, Don't Stop Pretraining: Adapt Language Models to Domains and TasksC//Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, P8342, DOI [DOI 10.18653/V1/2020.ACL-MAIN.740, 10.18653/v1/2020.aclmain.740]
[10]  
Han Xiaochuang, 2019, P 2019 C EMP METH NA, P4237