Enhancing relation extraction using multi-task learning with SDP evidence

被引:0
作者
Wang, Hailin [1 ,2 ]
Zhang, Dan [1 ,2 ]
Liu, Guisong [1 ,2 ]
Huang, Li [1 ,2 ]
Qin, Ke [3 ]
机构
[1] Southwestern Univ Finance & Econ, Sch Comp & Artificial Intelligence, Complex Lab New Finance & Econ, Chengdu 611130, Peoples R China
[2] Kash Inst Elect & Informat Ind, Kashgar, Xinjiang, Peoples R China
[3] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu, Sichuan, Peoples R China
基金
中国国家自然科学基金;
关键词
Relation extraction; Multi-task learning; Shortest dependency path; Evidence; ATTENTION; MODEL;
D O I
10.1016/j.ins.2024.120610
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Relation extraction (RE) is a crucial subtask of information extraction, which involves recognizing the relation between entity pairs in a sentence. Previous studies have extensively employed syntactic information, notably the shortest dependency path (SDP), to collect word evidence, termed SDP evidence, which gives clues about the given entity pair, thus improving RE. Nevertheless, prevalent transformer -based techniques lack syntactic information and cannot effectively model essential syntactic clues to support relations. This study exerts multi -task learning to address these issues by imbibing an SDP token position prediction task into the RE task. To this end, we introduce SGA, an SDP evidence guiding approach that transfers the SDP evidence into two novel supervisory signal labels: SDP tokens label and SDP matrix label. The former guides the attention modules to assign high attention weights to SDP token positions, emphasizing relational clues. In the meantime, the latter supervises SGA to predict a parameterized asymmetric product matrix among the SDP tokens for RE. Experimental outcomes demonstrate the model's enhanced ability to leverage SDP information, thereby directing attention modules and predicted matrix labels to focus on SDP evidence. Consequently, our proposed approach surpasses existing publicly available optimal baselines across four RE datasets: SemEval2010-Task8, KBP37, NYT, and WebNLG. 1
引用
收藏
页数:15
相关论文
共 45 条
  • [1] Soares LB, 2019, Arxiv, DOI [arXiv:1906.03158, 10.48550/arXiv.1906.03158]
  • [2] REEGAT: RoBERTa Entity Embedding and Graph Attention Networks Enhanced Sentence Representation for Relation Extraction
    Cai, Fengze
    Hu, Qiang
    Zhou, Renjie
    Xiong, Neal
    [J]. ELECTRONICS, 2023, 12 (11)
  • [3] Devlin J, 2019, Arxiv, DOI arXiv:1810.04805
  • [4] Cohen AD, 2021, Arxiv, DOI arXiv:2010.04829
  • [5] Creating Training Corpora for NLG Micro-Planning
    Gardent, Claire
    Shimorina, Anastasia
    Narayan, Shashi
    Perez-Beltrachini, Laura
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 179 - 188
  • [6] Semantic relation extraction using sequential and tree-structured LSTM with attention
    Geng, ZhiQiang
    Chen, GuoFei
    Han, YongMing
    Lu, Gang
    Li, Fang
    [J]. INFORMATION SCIENCES, 2020, 509 (509) : 183 - 192
  • [7] Guo Q., 2022, FINDINGS ASS COMPUTA, P1601
  • [8] Guo ZJ, 2020, Arxiv, DOI arXiv:1906.07510
  • [9] Hendrickx I., 2010, P 5 INT WORKSHOP SEM, P33
  • [10] Hu Xuming., 2023, Findings of the Association for Computa- tional Linguistics: ACL 2023, P10221, DOI DOI 10.18653/V1/2023.FINDINGS-ACL.649