MRE: A Military Relation Extraction Model Based on BiGRU and Multi-Head Attention

被引:11
作者
Lu, Yiwei [1 ]
Yang, Ruopeng [1 ]
Jiang, Xuping [1 ]
Zhou, Dan [1 ]
Yin, Changsheng [1 ]
Li, Zizhuo [2 ]
机构
[1] Natl Univ Def Technol, Coll Informat & Commun, Wuhan 430019, Peoples R China
[2] Wuhan Univ, Elect Informat Sch, Wuhan 430000, Peoples R China
来源
SYMMETRY-BASEL | 2021年 / 13卷 / 09期
关键词
military relation extraction; bi-directional encoder representations from transformers (BERT); BiGRU; multi-head attention; ENTITY;
D O I
10.3390/sym13091742
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
A great deal of operational information exists in the form of text. Therefore, extracting operational information from unstructured military text is of great significance for assisting command decision making and operations. Military relation extraction is one of the main tasks of military information extraction, which aims at identifying the relation between two named entities from unstructured military texts. However, the traditional methods of extracting military relations cannot easily resolve problems such as inadequate manual features and inaccurate Chinese word segmentation in military fields, failing to make full use of symmetrical entity relations in military texts. With our approach, based on the pre-trained language model, we present a Chinese military relation extraction method, which combines the bi-directional gate recurrent unit (BiGRU) and multi-head attention mechanism (MHATT). More specifically, the conceptual foundation of our method lies in constructing an embedding layer and combining word embedding with position embedding, based on the pre-trained language model; the output vectors of BiGRU neural networks are symmetrically spliced to learn the semantic features of context, and they fuse the multi-head attention mechanism to improve the ability of expressing semantic information. On the military text corpus that we have built, we conduct extensive experiments. We demonstrate the superiority of our method over the traditional non-attention model, attention model, and improved attention model, and the comprehensive evaluation value F1-score of the model is improved by about 4%.
引用
收藏
页数:15
相关论文
共 36 条
[1]  
Bunescu Razvan C., 2005, P HUMAN LANGUAGE TEC, P724, DOI DOI 10.3115/1220575.1220666
[2]  
[车万翔 Che Wanxiang], 2005, [中文信息学报, Journal of Chinese Information Processing], V19, P1
[3]  
Cho K., 2014, P C EMP METH NAT LAN, P1724, DOI DOI 10.3115/V1/D14-1179
[4]  
ChunYang Liu, 2013, Advanced Data Mining and Applications. 9th International Conference, ADMA 2013. Proceedings: LNCS 8347, P231, DOI 10.1007/978-3-642-53917-6_21
[5]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[6]   A Knowledge-Enriched and Span-Based Network for Joint Entity and Relation Extraction [J].
Ding, Kun ;
Liu, Shanshan ;
Zhang, Yuhao ;
Zhang, Hui ;
Zhang, Xiaoxiong ;
Wu, Tongtong ;
Zhou, Xiaolei .
CMC-COMPUTERS MATERIALS & CONTINUA, 2021, 68 (01) :377-389
[7]  
Doddington G., 2004, Lrec, P2014
[8]  
dos Santos CN, 2015, PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1, P626
[9]  
Girju R., 2007, Proceedings of the Fourth International Workshop on Semantic Evaluations (SemEval-2007), P13
[10]  
Graves A, 2012, STUD COMPUT INTELL, V385, P1, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]