Prompt-Based Prototypical Framework for Continual Relation Extraction

被引:10
作者
Zhang, Han [1 ,2 ,3 ]
Liang, Bin [1 ,4 ]
Yang, Min [5 ]
Wang, Hui [2 ]
Xu, Ruifeng [1 ,2 ,3 ]
机构
[1] Harbin Inst Technol Shenzhen, Sch Comp Sci & Technol, Shenzhen 518000, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518000, Peoples R China
[3] Guangdong Prov Key Lab Novel Secur Intelligence T, Shenzhen 518055, Peoples R China
[4] China Merchants Secur, Joint Lab HITSZ, Shenzhen 518000, Peoples R China
[5] Chinese Acad Sci, Shenzhen Inst Adv Technol, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Prototypes; Stability analysis; Feature extraction; Training; Speech processing; Security; Continual learning; relation extraction; catastrophic forgetting; prompt method; prototype;
D O I
10.1109/TASLP.2022.3199655
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Continual relation extraction (CRE) is an important task of continual learning, which aims to learn incessantly emerging new relations between entities from texts. To avoid catastrophically forgetting old relations, some existing research efforts have focused on exploring memory replayed methods by storing typical historical learned instances or embedding all observed relations as prototypes in the episodic memory and replaying them in the subsequent training process. However, they generally fail to exploit the relation knowledge contained in the pre-trained language model (PLM), which could provide enlightening information to the representations of new relations from the known ones. To this end, we investigate the CRE from a novel perspective by generating knowledge-infused relation prototypes to leverage the relational knowledge from PLM with prompt tuning. Specifically, based on the typical samples collected from the historical learned instances with K-means algorithm, we devise novel relational knowledge-infused prompts to elicit relational knowledge from PLM for generating knowledge-infused relation prototypes. Then the prototypes are used to refine the typical examples embedding and calculate the stability-plasticity balance score for adjusting the memory replayed progress. The experimental results show that our method outperforms the state-of-the-art baseline models in CRE. The further extensive analysis presents that the proposed method is robust to memory size, task order, length of the task sequence, and the number of training instances.
引用
收藏
页码:2801 / 2813
页数:13
相关论文
共 81 条
  • [41] Liu Y, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL) AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (IJCNLP), VOL 2, P285
  • [42] Lopez-Paz D, 2017, ADV NEUR IN, V30
  • [43] Mausam Mausam, 2016, P 25 INT JOINT C ART, P4074
  • [44] Min Bonan, 2012, EMNLP, P1027
  • [45] Moreira J., 2020, P INT JOINT C NEUR N, P1
  • [46] Nguyen C. V., 2018, INT C LEARN REPR ICL
  • [47] Nguyen T. H., 2015, P 1 WORKSH VECT SPAC, P39
  • [48] Obamuyide A, 2019, 4TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP (REPL4NLP-2019), P224
  • [49] Continual lifelong learning with neural networks: A review
    Parisi, German I.
    Kemker, Ronald
    Part, Jose L.
    Kanan, Christopher
    Wermter, Stefan
    [J]. NEURAL NETWORKS, 2019, 113 : 54 - 71
  • [50] Pennington Jeffrey, 2014, P 2014 C EMPIRICAL M, P1532, DOI [10.3115/v1/D14-1162, DOI 10.3115/V1/D14-1162]