A Graph Convolutional Network With Multiple Dependency Representations for Relation Extraction

被引:16
作者
Hu, Yanfeng [1 ,2 ]
Shen, Hong [1 ,2 ]
Liu, Wuling [1 ,2 ]
Min, Fei [1 ,2 ]
Qiao, Xue [1 ,2 ]
Jin, Kangrong [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Elect, Suzhou 215123, Peoples R China
[2] Key Lab Intelligent Aerosp Big Data Applicat Tech, Suzhou 215123, Peoples R China
来源
IEEE ACCESS | 2021年 / 9卷
关键词
Task analysis; Computational modeling; Semantics; Analytical models; Syntactics; Natural language processing; Feature extraction; Graph convolutional network; relation extraction; syntactic dependency tree;
D O I
10.1109/ACCESS.2021.3086480
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Dependency analysis can assist neural networks to capture semantic features within a sentence for entity relation extraction (RE). Both hard and soft strategies of encoding dependency tree structure have been developed to balance the beneficial extra information against the unfavorable interference in the task of RE. A wide application of graph convolutional network (GCN) in the field of natural language processing (NLP) has demonstrated its effectiveness in encoding the input sentence with the dependency tree structure, as well as its efficiency in parallel computation. This study proposes a novel GCN-based model using multiple representations to depict the dependency tree from various perspectives, and combines those dependency representations afterward to obtain a better sentence representation for relation classification. This model can maximally draw from the sentence the semantic features relevant to the relationship between entities. Results show that our model achieves state-of-the-art performance in terms of the F-1 score (68.0) on the Text Analysis Conference relation extraction dataset (TACRED). In addition, we verify that the renormalization parameter in the GCN operation should be carefully chosen to help GCN-based models achieve its best performance.
引用
收藏
页码:81575 / 81587
页数:13
相关论文
共 55 条
  • [1] Ando Rie Kubota, 2006, Adv Neural Inf Process Syst
  • [2] [Anonymous], 2009, P 5 WORKSH SEM EV RE
  • [3] Bruna J., 2013, SPECTRAL NETWORKS LO
  • [4] Bunescu Razvan C., 2005, P HUMAN LANGUAGE TEC, P724, DOI DOI 10.3115/1220575.1220666
  • [5] Cai R, 2016, PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P756
  • [6] Can DC, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P2902
  • [7] Chung F. R., 1997, Regional Conference Series in Mathematics
  • [8] Culotta A., 2004, P 42 ANN M ASS COMP, P423, DOI [DOI 10.3115/1218955.1219009, /10.3115/1218955.1219009]
  • [9] Dat P., 2007, Proceedings of A AAI-07, P1414
  • [10] Defferrard M, 2016, ADV NEUR IN, V29