REEGAT: RoBERTa Entity Embedding and Graph Attention Networks Enhanced Sentence Representation for Relation Extraction

被引:2
作者
Cai, Fengze [1 ]
Hu, Qiang [2 ]
Zhou, Renjie [2 ]
Xiong, Neal [3 ]
机构
[1] Hangzhou Dianzi Univ, Zhuoyue Honors Coll, Hangzhou 310018, Peoples R China
[2] Hangzhou Dianzi Univ, Coll Comp Sci & Technol, Hangzhou 310018, Peoples R China
[3] Sul Ross State Univ, Dept Comp Math & Phys Sci, Alpine, TX 79830 USA
关键词
information extraction; relation extraction; contrastive learning; graph attention network; INFORMATION EXTRACTION;
D O I
10.3390/electronics12112429
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Relation extraction is one of the most important intelligent information extraction technologies, which can be used to construct and optimize services in intelligent communication systems (ICS). One issue with the existing relation extraction approaches is that they use one-sided sentence embedding as their final prediction vector, which degrades relation extraction performance. The innovative relation extraction model REEGAT (RoBERTa Entity Embedding and Graph Attention networks enhanced sentence representation) that we present in this paper, incorporates the concept of enhanced word embedding from graph neural networks. The model first uses RoBERTa to obtain word embedding and PyTorch embedding to obtain relation embedding. Then, the multi-headed attention mechanism in GAT (graph attention network) is introduced to weight the word embedding and relation embedding to enrich further the meaning conveyed by the word embedding. Finally, the entity embedding component is used to obtain sentence representation by pooling the word embedding from GAT and the entity embedding from named entity recognition. The weighted and pooled word embedding contains more relational information to alleviate the one-sided problem of sentence representation. The experimental findings demonstrate that our model outperforms other standard methods.
引用
收藏
页数:16
相关论文
共 37 条
[1]   A deep learning based trust- and tag-aware recommender system [J].
Ahmadian, Sajad ;
Ahmadian, Milad ;
Jalili, Mahdi .
NEUROCOMPUTING, 2022, 488 :557-571
[2]  
Arora S., 2023, KNOWLEDGE RETRIEVAL
[3]  
Atkinson M., 2009, User Centric Media, P321
[4]   KNN-BLOCK DBSCAN: Fast Clustering for Large-Scale Data [J].
Chen, Yewang ;
Zhou, Lida ;
Pei, Songwen ;
Yu, Zhiwen ;
Chen, Yi ;
Liu, Xin ;
Du, Jixiang ;
Xiong, Naixue .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2021, 51 (06) :3939-3953
[5]  
Chen YL, 2020, ADV SOC SCI EDUC HUM, V471, P198
[6]   Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks [J].
Chiang, Wei-Lin ;
Liu, Xuanqing ;
Si, Si ;
Li, Yang ;
Bengio, Samy ;
Hsieh, Cho-Jui .
KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, :257-266
[7]   TERRORISM INFORMATION EXTRACTION FROM ONLINE REPORTS [J].
Conlon, Sumali J. ;
Abrahams, Alan S. ;
Simmons, Lakisha L. .
JOURNAL OF COMPUTER INFORMATION SYSTEMS, 2015, 55 (03) :20-28
[8]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[9]  
Giorgi J, 2019, Arxiv, DOI arXiv:1912.13415
[10]  
Guo ZJ, 2020, Arxiv, DOI arXiv:1906.07510