CONSPROMPT: EXPLOITING CONTRASTIVE SAMPLES FOR FEW-SHOT PROMPT LEARNING

被引:0
作者
Weng, Jinta [1 ,2 ]
Deng, Yifan [1 ,2 ]
Li, Donghao [1 ,2 ]
You, Hao [1 ,2 ]
Hu, Yue [1 ,2 ]
Huang, Heyan [3 ,4 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[3] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing, Peoples R China
[4] Southeast Acad Informat Technol, Putian, Peoples R China
来源
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024 | 2024年
基金
中国国家自然科学基金;
关键词
Prompt learning; Pre-trained language model; contrastive learning; few-shot learning;
D O I
10.1109/ICASSP48485.2024.10448403
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Prompt has become an effective linguistic tool for utilizing pre-trained language models. However, in few-shot scenarios, subtle changes of prompt's design always make the result widely different, and the prompt learning methods are also easy to overfit the limited samples. To alleviate this, we explore utilizing suitable contrastive samples and multi-degree contrastive learning methods to improve the robustness of prompt's representation. Therefore, the proposed Consprompt combined with prompt encoding network, contrastive sampling modules, and contrastive scoring modules, is introduced to realize differential contrastive learning. Our results exhibit the state-of-the-art performance in different few-shot settings, and the ablation experiments also certify the effectiveness of utilizing multi-degree contrastive learning in prompt-based fine-tuning process.
引用
收藏
页码:6835 / 6839
页数:5
相关论文
共 16 条
[1]  
[Anonymous], 2023, INT C COMP SCI
[2]  
Brown TB, 2020, ADV NEUR IN, V33
[3]  
Chen T., 2020, INT C MACH LEARN PML, P1597
[4]  
Gao TY, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, P3816
[5]  
Jian YR, 2022, NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, P5577
[6]  
Li XLS, 2021, 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, P4582
[7]  
Liu X., 2021, arXiv
[8]  
Liu Yang., 2019, INFORM SYST RES, DOI DOI 10.48550/ARXIV.1907.11692
[9]  
Madotto Andrea, 2021, arXiv
[10]  
Radford Alec., 2019, Language Models are Unsupervised Multitask Learners, V1, P9