Language Model Pre-training Method in Machine Translation Based on Named Entity Recognition

被引:8
作者
Li, Zhen [1 ]
Qu, Dan [1 ]
Xie, Chaojie [2 ]
Zhang, Wenlin [1 ]
Li, Yanxia [3 ]
机构
[1] PLA Strateg Support Force Informat Engn Univ, Informat Syst Engn Coll, 93 Hightech Zone, Zhengzhou 450000, Peoples R China
[2] Zhengzhou Xinda Inst Adv Technol, 93 Hightech Zone, Zhengzhou 450000, Peoples R China
[3] PLA Strateg Support Force Informat Engn Univ, Foreign Languages Coll, 93 Hightech Zone, Zhengzhou 450000, Peoples R China
关键词
Unsupervised machine translation; language model; named entity recognition;
D O I
10.1142/S0218213020400217
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural Machine Translation (NMT) model has become the mainstream technology in machine translation. The supervised neural machine translation model trains with abundant of sentence-level parallel corpora. But for low-resources language or dialect with no such corpus available, it is difficult to achieve good performance. Researchers began to focus on unsupervised neural machine translation (UNMT) that monolingual corpus as training data. UNMT need to construct the language model (LM) which learns semantic information from the monolingual corpus. This paper focuses on the pre-training of LM in unsupervised machine translation and proposes a pre-training method, NER-MLM (named entity recognition masked language model). Through performing NER, the proposed method can obtain better semantic information and language model parameters with better training results. In the unsupervised machine translation task, the BLEU scores on the WMT'16 English-French, English-German, data sets are 35.30, 27.30 respectively. To the best of our knowledge, this is the highest results in the field of UNMT reported so far.
引用
收藏
页数:10
相关论文
共 50 条
  • [41] Military Named Entity Recognition Method Based on Deep Learning
    Wang, Xuefeng
    Yang, Ruopeng
    Lu, Yiwei
    Wu, Qingfeng
    PROCEEDINGS OF 2018 5TH IEEE INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND INTELLIGENCE SYSTEMS (CCIS), 2018, : 479 - 483
  • [42] Improving Norwegian Translation of Bicycle Terminology Using Custom Named-Entity Recognition and Neural Machine Translation
    Hellebust, Daniel
    Lawal, Isah A.
    ELECTRONICS, 2023, 12 (10)
  • [43] An effective undersampling method for biomedical named entity recognition using machine learning
    Archana, S. M.
    Prakash, Jay
    EVOLVING SYSTEMS, 2024, 15 (04) : 1541 - 1549
  • [44] A novel large-language-model-driven framework for named entity recognition
    Wang, Zhenhua
    Chen, Huiru
    Xu, Guang
    Ren, Ming
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (03)
  • [45] Chinese mineral named entity recognition based on BERT model
    Yu, Yuqing
    Wang, Yuzhu
    Mua, Jingqin
    Li, Wei
    Jiao, Shoutao
    Wang, Zhenhua
    Lv, Pengfei
    Zhu, Yueqin
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 206
  • [46] Enhancing named entity recognition with external knowledge from large language model
    Li, Qi
    Xie, Tingyu
    Zhang, Jian
    Ma, Ke
    Su, Jiayuan
    Yang, Kaixiang
    Wang, Hongwei
    KNOWLEDGE-BASED SYSTEMS, 2025, 318
  • [47] Maximum Entropy Named Entity Recognition for Czech Language
    Konkol, Michal
    Konopik, Miloslav
    TEXT, SPEECH AND DIALOGUE, TSD 2011, 2011, 6836 : 203 - 210
  • [48] Large Language Models for Latvian Named Entity Recognition
    Viksna, Rinalds
    Skadina, Inguna
    HUMAN LANGUAGE TECHNOLOGIES - THE BALTIC PERSPECTIVE (HLT 2020), 2020, 328 : 62 - 69
  • [49] Deep learning with language models improves named entity recognition for PharmaCoNER
    Sun, Cong
    Yang, Zhihao
    Wang, Lei
    Zhang, Yin
    Lin, Hongfei
    Wang, Jian
    BMC BIOINFORMATICS, 2021, 22 (SUPPL 1)
  • [50] Deep learning with language models improves named entity recognition for PharmaCoNER
    Cong Sun
    Zhihao Yang
    Lei Wang
    Yin Zhang
    Hongfei Lin
    Jian Wang
    BMC Bioinformatics, 22