FPrompt-PLM: Flexible-Prompt on Pretrained Language Model for Continual Few-Shot Relation Extraction

被引:0
作者
Zhang, Lingling [1 ,2 ]
Li, Yifei [1 ,2 ]
Wang, Qianying [3 ]
Wang, Yun [4 ,5 ]
Yan, Hang [1 ,2 ]
Wang, Jiaxin [6 ,7 ,8 ]
Liu, Jun [6 ,7 ,8 ]
机构
[1] Xi An Jiao Tong Univ, Sch Comp Sci & Engn, Key Lab Intelligent Networks & Network Secur, Xian 710049, Peoples R China
[2] Xi An Jiao Tong Univ, Minist Educ, Key Lab Intelligent Networks & Network Secur, Xian, Peoples R China
[3] Lenovo Res, Beijing 100085, Peoples R China
[4] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100190, Peoples R China
[5] China Mobile Informat Syst Integrat Co Ltd, Beijing 100032, Peoples R China
[6] Xi An Jiao Tong Univ, Sch Comp Sci & Technol, Xian 710049, Peoples R China
[7] Xi An Jiao Tong Univ, Shaanxi Prov Key Lab Big Data Knowledge Engn, Xian 710049, Peoples R China
[8] SHMEEA, BigKE Joint Innovat Ctr, Shanghai 200082, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Prototypes; Training; Testing; Adaptation models; Semantics; Computational modeling; Relation extraction; continual few-shot learning; flexible-prompt; continual meta-finetuning; prototype diversity; NETWORK;
D O I
10.1109/TKDE.2024.3419117
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Relation extraction (RE) aims to identify the relation between two entities within a sentence, which plays a crucial role in information extraction. Traditional supervised setting on RE does not fit the actual scenario, due to the continuous emergence of new relations and the unavailability of massive labeled examples. Continual few-shot relation extraction (CFS-RE) is proposed as a potential solution to the above situation, which requires the model to learn new relations sequentially from a few examples. Apparently, CFS-RE is more challenging than previous RE, as the catastrophic forgetting of old knowledge and few-shot overfitting on a handful of examples. To this end, we propose a novel flexible-prompt framework on pretrained language model named FPrompt-PLM for CFS-RE, which includes flexible-prompt embedding, pretrained-language understanding, and nearest-prototype learning modules. Note that two pools in FPrompt-PLM, i.e., prompt and prototype pools, are continual updated and applied for prediction of all seen relations at current time-step. The former pool records the distinctive prompt embedding in each time period, and the latter records all learned relation prototypes. Besides, three progressive stages are introduced to learn FPrompt-PLM's parameters and apply this model for CFS-RE testing, which includes meta-training, continual meta-finetuning, and testing stages. And we improve the CFS-RE loss by incorporating multiple distillation losses as well as a novel prototype-diversity loss in these stages to alleviate the catastrophic forgetting and few-shot overfitting problems. Comprehensive experiments on two widely-used datasets show that FPrompt-PLM achieves significant performance improvements over the SOTA baselines.
引用
收藏
页码:8267 / 8282
页数:16
相关论文
共 8 条
  • [1] A prompt tuning method based on relation graphs for few-shot relation extraction
    Zhang, Zirui
    Yang, Yiyu
    Chen, Benhui
    NEURAL NETWORKS, 2025, 185
  • [2] Joint data augmentation and knowledge distillation for few-shot continual relation extraction
    Zhongcheng Wei
    Yunping Zhang
    Bin Lian
    Yongjian Fan
    Jijun Zhao
    Applied Intelligence, 2024, 54 : 3516 - 3528
  • [3] Joint data augmentation and knowledge distillation for few-shot continual relation extraction
    Wei, Zhongcheng
    Zhang, Yunping
    Lian, Bin
    Fan, Yongjian
    Zhao, Jijun
    APPLIED INTELLIGENCE, 2024, 54 (04) : 3516 - 3528
  • [4] Knowledge-enhanced meta-prompt for few-shot relation extraction
    Cui, Jinman
    Xu, Fu
    Wang, Xinyang
    Li, Yakun
    Qu, Xiaolong
    Yao, Lei
    Li, Dongmei
    COMPUTER SPEECH AND LANGUAGE, 2025, 91
  • [5] Few-Shot Relation Extraction Through Prompt With Relation Information and Multi-Level Contrastive Learning
    Dong, Ye
    Yang, Rong
    Liu, Junbao
    Qin, Xizhong
    IEEE ACCESS, 2024, 12 : 123352 - 123361
  • [6] Rehearsal-free continual few-shot relation extraction via contrastive weighted prompts
    Yang, Fengqin
    Ren, Mengen
    Kong, Delu
    Liu, Shuhua
    Fu, Zhiguo
    NEUROCOMPUTING, 2025, 633
  • [7] An angular shrinkage BERT model for few-shot relation extraction with none-of-the-above detection
    Wang, Junwen
    Gao, Yongbin
    Fang, Zhijun
    PATTERN RECOGNITION LETTERS, 2023, 166 : 151 - 158
  • [8] CoCoOpter: Pre-train, prompt, and fine-tune the vision-language model for few-shot image classification
    Yan, Jie
    Xie, Yuxiang
    Guo, Yanming
    Wei, Yingmei
    Zhang, Xiaoping
    Luan, Xidao
    INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2023, 12 (02)