Sequence to sequence learning for joint extraction of entities and relations

被引:8
作者
Liang, Zeyu [1 ]
Du, Junping [2 ]
机构
[1] Commun Univ China, Sch Data Sci & Intelligent Media, Beijing, Peoples R China
[2] Beijing Univ Posts & Telecommun, Sch Comp Sci, Beijing Key Lab Intelligent Telecommun Software &, Beijing 100876, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Sequence to sequence learning; Gate graph neural networks; Joint extraction of entities and relations;
D O I
10.1016/j.neucom.2022.05.074
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The relations of two entities contained in the sentence are often complicated. There exists multiple relation tuples which owns one or both the same entities among them. Extracting those overlapped multiple relation tuples faces great challenges. Most existing works employ the famous Seq2Seq model for joint extraction of entities and relations, which cannot handle the multiple relations properly due to the missing of graph structure information of the relations and entities in a sentence. In this paper, we propose Seq2Seq-RE, an end-to-end relation extraction model, which first utilizes the gate graph neural networks (GGNNs) for joint extraction of entities and relations. Unlike previous works, we take the interaction of entities and relations through a GGNNs-based sequence-to-sequence with attention mechanism for better extracting multiple relation tuples. Besides, our graph structure based on the forward-edge, backward-edge, self-edge, and dependency-edge. To our knowledge, we are the first to employ four types of edge into graph-based neural networks for joint extraction of entities and relations and conduct a comprehensive qualitative analysis to investigate what types of edges benefit our task. Experimental results show that our model surpasses the current strong baseline methods by 1.7% in NYT29 and 0.8% in NYT24 (F1 score). (c) 2022 Published by Elsevier B.V.
引用
收藏
页码:480 / 488
页数:9
相关论文
共 33 条
  • [1] [Anonymous], 2007, Introduction to Statistical Relational Learning
  • [2] Bahdanau D., 2014, CORR, DOI DOI 10.48550/ARXIV.1409.0473
  • [3] Beck D, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, P273
  • [4] Brockschmidt R., 2016, P ICLR, P3104
  • [5] Chan Y.S., 2011, P 49 ANN M ASS COMP, P551
  • [6] Das R., 2018, P ICLR
  • [7] Fu TJ, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P1409
  • [8] Guo ZJ, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P241
  • [9] Hendrickx I., 2019, ARXIV, DOI DOI 10.48550/ARXIV.1911.10422
  • [10] Katiyar A, 2016, PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P919