NMT Enhancement based on Knowledge Graph Mining with Pre-trained Language Model

被引:0
作者
Yang, Hao [1 ]
Qin, Ying [1 ]
Deng, Yao [1 ]
Wang, Minghan [1 ]
机构
[1] Huawei Co Ltd, Translate Serv Ctr, Beijing, Peoples R China
来源
2020 22ND INTERNATIONAL CONFERENCE ON ADVANCED COMMUNICATION TECHNOLOGY (ICACT): DIGITAL SECURITY GLOBAL AGENDA FOR SAFE SOCIETY! | 2020年
关键词
NMT; Pre-trained Language Model; Knowledge Graph;
D O I
10.23919/icact48636.2020.9061292
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Pre-trained language models like Bert, RoBERT a, GPT, etc. have achieved SOTA effects on multiple NLP tasks (e.g. sentiment classification, information extraction, event extraction, etc.). We propose a simple method based on knowledge graph to improve the quality of machine translation. First, we propose a multi-task learning model that learns subjects, objects, and predicates at the same time. Second, we treat different predicates as different fields, and improve the recognition ability of NMT models in different fields through classification labels. Finally, beam search combined with L2R, R2L rearranges results through entities. Based on the CWMT2018 experimental data, using the predicate's domain classification identifier, the BLUE score increased from 33.58% to 37.63%, and through L2R, R2L rearrangement, the BLEU score increased to 39.25%, overall improvement is more than 5 percentage
引用
收藏
页码:185 / 189
页数:5
相关论文
共 19 条
[1]  
Alberti Chris, 2019, ABS190108634 ARXIV
[2]  
[Anonymous], 2018, ARXIV181210860
[3]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[4]  
Gupta S., 2019, ARXIV190801767
[5]  
Huang W., 2019, ARXIV190805908
[6]   Log-Linear Model Based Behavior Selection Method for Artificial Fish Swarm Algorithm [J].
Huang, Zhehuang ;
Chen, Yidong .
COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2015, 2015
[7]  
Jm D., 2019, ARXIV190711932
[8]  
Liu Zhibin, 2019, ARXIV190310245
[9]  
Nangia N., 2019, ARXIV190510425
[10]  
Peng W, 2019, FOURTH CONFERENCE ON MACHINE TRANSLATION (WMT 2019), VOL 3: SHARED TASK PAPERS, DAY 2, P164