Enhancing Document-Level Relation Extraction with Attention-Convolutional Hybrid Networks and Evidence Extraction

被引:0
作者
Zhang, Feiyu [1 ,2 ]
Hu, Ruiming [1 ,2 ]
Duan, Guiduo [2 ,3 ]
Huang, Tianxi [4 ]
机构
[1] Univ Elect Sci & Technol China, Shenzhen Inst Adv Study, Shenzhen, Peoples R China
[2] Univ Elect Sci & Technol China, Lab Intelligent Collaborat Comp, Chengdu, Peoples R China
[3] Trusted Cloud Comp & Big Data Key Lab Sichuan Prov, Chengdu, Peoples R China
[4] Chengdu Text Coll, Dept Fundamental Courses, Chengdu, Peoples R China
基金
中国国家自然科学基金;
关键词
Document-level relation extraction; Multi-head attention; Hybrid networks;
D O I
10.1007/s12559-024-10269-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Document-level relation extraction aims at extracting relations between entities in a document. In contrast to sentence-level correspondences, document-level relation extraction requires reasoning over multiple sentences to extract complex relational triples. Recent work has found that adding additional evidence extraction tasks and using the extracted evidence sentences to help predict can improve the performance of document-level relation extraction tasks, however, these approaches still face the problem of inadequate modeling of the interactions between entity pairs. In this paper, based on the review of human cognitive processes, we propose a hybrid network HIMAC applied to the entity pair feature matrix, in which the multi-head attention sub-module can fuse global entity-pair information on a specific inference path, while the convolution sub-module is able to obtain local information of adjacent entity pairs. Then we incorporate the contextual interaction information learned by the entity pairs into the relation prediction and evidence extraction tasks. Finally, the extracted evidence sentences are used to further correct the relation extraction results. We conduct extensive experiments on two document-level relation extraction benchmark datasets (DocRED/Re-DocRED), and the experimental results demonstrate that our method achieves state-of-the-art performance (62.84/75.89 F1). Experiments demonstrate the effectiveness of the proposed method.
引用
收藏
页码:1113 / 1124
页数:12
相关论文
共 40 条
[1]  
Cheng Q, 2021, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, P2819
[2]  
Christopoulou F, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P4925
[3]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[4]  
Goyal P, 2018, Arxiv, DOI [arXiv:1706.02677, DOI 10.48550/ARXIV.1706.02677]
[5]  
Guo ZJ, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P241
[6]  
Huang K, 2021, REPL4NLP 2021: PROCEEDINGS OF THE 6TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP, P307
[7]  
Jia R, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P3693
[8]  
Jiang Feng, 2022, P 29 INT C COMP LING, P1904
[9]  
Li B., 2020, P 28 INT C COMP LING, P1551
[10]  
Loshchilov I., 2019, INT C LEARNING REPRE