Evidence and Axial Attention Guided Document-level Relation Extraction

被引:0
作者
Yuan, Jiawei [1 ,3 ,4 ]
Leng, Hongyong [3 ,4 ,5 ]
Qian, Yurong [1 ,3 ,4 ]
Chen, Jiaying [1 ,3 ,4 ]
Ma, Mengnan [1 ,3 ,4 ]
Hou, Shuxiang [2 ]
机构
[1] Xinjiang Univ, Sch Software, Urumqi 830000, Peoples R China
[2] Xinjiang Univ, Sch Informat Sci & Engn, Urumqi, Peoples R China
[3] Key Lab Signal Detect & Proc Xinjiang Uygur Autono, Urumqi, Peoples R China
[4] Xinjiang Univ, Key Lab Software Engn, Urumqi, Peoples R China
[5] Beijing Inst Technol, Sch Comp Sci & Technol, BIT, Beijing 100081, Peoples R China
基金
中国国家自然科学基金;
关键词
Information extraction; Document-level Relation Extraction; Evidence retrieval; Axial attention;
D O I
10.1016/j.csl.2024.101728
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Document-level Relation Extraction (DocRE) aims to identify semantic relations among multiple entity pairs within a document. Most of the previous DocRE methods take the entire document as input. However, for human annotators, a small subset of sentences in the document, namely the evidence, is sufficient to infer the relation of an entity pair. Additionally, a document usually contains multiple entities, and these entities are scattered throughout various location of the document. Previous models use these entities independently, ignore the global interdependency among relation triples. To handle above issues, we propose a novel framework EAAGRE (Evidence and Axial Attention Guided Relation Extraction). Firstly, we use human-annotated evidence labels to supervise the attention module of DocRE system, making the model pay attention to the evidence sentences rather than others. Secondly, we construct an entity-level relation matrix and use axial attention to capture the global interactions among entity pairs. By doing so, we further extract the relations that require multiple entity pairs for prediction. We conduct various experiments on DocRED and have some improvement compared to baseline models, verifying the effectiveness of our model.
引用
收藏
页数:11
相关论文
共 34 条
[1]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[2]  
Huang K, 2021, REPL4NLP 2021: PROCEEDINGS OF THE 6TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP, P307
[3]  
Huang QZ, 2021, ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, P998
[4]  
Ilya L, 2018, INT C LEARN REPR
[5]  
Jia R, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P3693
[6]  
Lin YK, 2016, PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P2124
[7]  
Liu H., 2023, ADV KNOWLEDGE DISCOV, P316, DOI [10.1007/978-3-031, DOI 10.1007/978-3-031]
[8]  
Liu YH, 2019, Arxiv, DOI [arXiv:1907.11692, 10.48550/arXiv.1907.11692, DOI 10.48550/ARXIV.1907.11692]
[9]  
Ma YM, 2023, 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, P1971
[10]  
Mintz M, 2009, P JOINT C 47 ANN M A, P1003, DOI DOI 10.3115/1690219.1690287