Prompt template construction by Average Gradient Search with External Knowledge for aspect sentimental analysis

被引:8
作者
Du, Yongping [1 ]
Yin, Zihao [1 ]
Xie, Runfeng [1 ]
Zhang, Qi [1 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Aspect-based Sentiment Analysis; Prompt learning; Pre-trained language model; Natural Language Processing; Average Gradient Search; External knowledge;
D O I
10.1016/j.eswa.2023.122271
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Aspect-based Sentiment Analysis (ABSA) aims to predict the sentiment polarity towards a particular aspect in a sentence. Most of the existing methods construct neural networks or fine-tune the pre-trained language models. However, it is difficult to fully utilize the knowledge of language model learned during pretraining process, and it usually performs worse in few-shot experiment. Prompt learning can alleviate the above problems effectively. Manual prompt template construction methods are adopted usually but it leads to high costs and low efficiency, and there are seldom prompt templates applicable to ABSA task. We propose a method for ABSA prompt template construction, named Average Gradient Search with External Knowledge Based on KNN. The vocabulary list of the pre-trained language model is built as a KD-Tree, and the KNN algorithm is used to search the best prompt template on KD-Tree. For the sentiment polarity prediction, the best prompt template is used to wrap the ABSA data, converting the classification task to generative task, which enables the language model to fully leverage the learned knowledge from the pre-training stage. Moreover, the constructed verbalizer incorporates external knowledge to provide label words of each class with extensive semantic coverage and lower subjective bias. The comparison experimental results on SemEval2014 Restaurant and Laptop datasets demonstrate that the proposed method outperforms the existing SOTA models. Further, the few-shot experimental results indicate that, with only a small number of training samples, the proposed method achieves comparable or even better performance than the baseline model trained on the full dataset.
引用
收藏
页数:9
相关论文
共 37 条
[1]  
Brown TB, 2020, Arxiv, DOI [arXiv:2005.14165, DOI 10.48550/ARXIV.2005.14165]
[2]  
Cai HJ, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), P340
[3]  
Chen H, 2022, PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), P2974
[4]   KnowPrompt: Knowledge-aware Prompt-tuning with Synergistic Optimization for Relation Extraction [J].
Chen, Xiang ;
Zhang, Ningyu ;
Xie, Xin ;
Deng, Shumin ;
Yao, Yunzhi ;
Tan, Chuanqi ;
Huang, Fei ;
Si, Luo ;
Chen, Huajun .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :2778-2788
[5]  
Chen Y., 2022, P 2022 C EMPIRICAL M, P4300
[6]  
Deng M., 2022, P 2022 C EMPIRICAL M, P3369
[7]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[8]  
Fan FF, 2018, 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), P3433
[9]  
Gao TY, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, P3816
[10]  
Jiang QN, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P6280