Prompt-Based Prototypical Framework for Continual Relation Extraction

被引:10
作者
Zhang, Han [1 ,2 ,3 ]
Liang, Bin [1 ,4 ]
Yang, Min [5 ]
Wang, Hui [2 ]
Xu, Ruifeng [1 ,2 ,3 ]
机构
[1] Harbin Inst Technol Shenzhen, Sch Comp Sci & Technol, Shenzhen 518000, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518000, Peoples R China
[3] Guangdong Prov Key Lab Novel Secur Intelligence T, Shenzhen 518055, Peoples R China
[4] China Merchants Secur, Joint Lab HITSZ, Shenzhen 518000, Peoples R China
[5] Chinese Acad Sci, Shenzhen Inst Adv Technol, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Prototypes; Stability analysis; Feature extraction; Training; Speech processing; Security; Continual learning; relation extraction; catastrophic forgetting; prompt method; prototype;
D O I
10.1109/TASLP.2022.3199655
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Continual relation extraction (CRE) is an important task of continual learning, which aims to learn incessantly emerging new relations between entities from texts. To avoid catastrophically forgetting old relations, some existing research efforts have focused on exploring memory replayed methods by storing typical historical learned instances or embedding all observed relations as prototypes in the episodic memory and replaying them in the subsequent training process. However, they generally fail to exploit the relation knowledge contained in the pre-trained language model (PLM), which could provide enlightening information to the representations of new relations from the known ones. To this end, we investigate the CRE from a novel perspective by generating knowledge-infused relation prototypes to leverage the relational knowledge from PLM with prompt tuning. Specifically, based on the typical samples collected from the historical learned instances with K-means algorithm, we devise novel relational knowledge-infused prompts to elicit relational knowledge from PLM for generating knowledge-infused relation prototypes. Then the prototypes are used to refine the typical examples embedding and calculate the stability-plasticity balance score for adjusting the memory replayed progress. The experimental results show that our method outperforms the state-of-the-art baseline models in CRE. The further extensive analysis presents that the proposed method is robust to memory size, task order, length of the task sequence, and the number of training instances.
引用
收藏
页码:2801 / 2813
页数:13
相关论文
共 81 条
  • [1] Rusu AA, 2016, Arxiv, DOI arXiv:1606.04671
  • [2] Aljundi R, 2019, ADV NEUR IN, V32
  • [3] Memory Aware Synapses: Learning What (not) to Forget
    Aljundi, Rahaf
    Babiloni, Francesca
    Elhoseiny, Mohamed
    Rohrbach, Marcus
    Tuytelaars, Tinne
    [J]. COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 : 144 - 161
  • [4] [Anonymous], 2018, Advances in neural information processing systems
  • [5] Pattern-based approaches to semantic relation extraction -: A state-of-the-art
    Auger, Alain
    Barriere, Caroline
    [J]. TERMINOLOGY, 2008, 14 (01): : 1 - 19
  • [6] RECON: Relation Extraction using Knowledge Graph Context in a Graph Neural Network
    Bastos, Anson
    Nadgeri, Abhishek
    Singh, Kuldeep
    Mulang, Isaiah Onando
    Shekarpour, Saeedeh
    Hoffart, Johannes
    Kaul, Manohar
    [J]. PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2021 (WWW 2021), 2021, : 1673 - 1685
  • [7] Bengio Y., 2014, P INT C LEARN REPR B
  • [8] Biesialska M. M., 2020, P 28 INT C COMPUTATI, P6523, DOI 10.18653/v1/2020.coling-main.574
  • [9] Bordes A, 2015, Arxiv, DOI arXiv:1506.02075
  • [10] Chaudhry A., 2018, P INT C LEARN REPR I