Language Model Pre-training Method in Machine Translation Based on Named Entity Recognition

被引:8
作者
Li, Zhen [1 ]
Qu, Dan [1 ]
Xie, Chaojie [2 ]
Zhang, Wenlin [1 ]
Li, Yanxia [3 ]
机构
[1] PLA Strateg Support Force Informat Engn Univ, Informat Syst Engn Coll, 93 Hightech Zone, Zhengzhou 450000, Peoples R China
[2] Zhengzhou Xinda Inst Adv Technol, 93 Hightech Zone, Zhengzhou 450000, Peoples R China
[3] PLA Strateg Support Force Informat Engn Univ, Foreign Languages Coll, 93 Hightech Zone, Zhengzhou 450000, Peoples R China
关键词
Unsupervised machine translation; language model; named entity recognition;
D O I
10.1142/S0218213020400217
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural Machine Translation (NMT) model has become the mainstream technology in machine translation. The supervised neural machine translation model trains with abundant of sentence-level parallel corpora. But for low-resources language or dialect with no such corpus available, it is difficult to achieve good performance. Researchers began to focus on unsupervised neural machine translation (UNMT) that monolingual corpus as training data. UNMT need to construct the language model (LM) which learns semantic information from the monolingual corpus. This paper focuses on the pre-training of LM in unsupervised machine translation and proposes a pre-training method, NER-MLM (named entity recognition masked language model). Through performing NER, the proposed method can obtain better semantic information and language model parameters with better training results. In the unsupervised machine translation task, the BLEU scores on the WMT'16 English-French, English-German, data sets are 35.30, 27.30 respectively. To the best of our knowledge, this is the highest results in the field of UNMT reported so far.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Supervised Named Entity Recognition in Assamese language
    Talukdar, Gitimoni
    Borah, Pranjal Protim
    Baruah, Arup
    2014 INTERNATIONAL CONFERENCE ON CONTEMPORARY COMPUTING AND INFORMATICS (IC3I), 2014, : 187 - 191
  • [32] A Method of Named Entity Recognition for Tigrinya
    Yohannes, Hailemariam Mehari
    Amagasa, Toshiyuki
    APPLIED COMPUTING REVIEW, 2022, 22 (03): : 56 - 68
  • [33] FEATURES FOR NAMED ENTITY RECOGNITION IN CZECH LANGUAGE
    Kral, Pavel
    KEOD 2011: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON KNOWLEDGE ENGINEERING AND ONTOLOGY DEVELOPMENT, 2011, : 437 - 441
  • [34] Incorporating Class-based Language Model for Named Entity Recognition in Factorized Neural Transducer
    Wang, Peng
    Yang, Yifan
    Bang, Zheng
    Tan, Tian
    Zhang, Shiliang
    Chen, Xie
    INTERSPEECH 2024, 2024, : 742 - 746
  • [35] Named entity recognition in Odia language: a rule-based approach
    Anandika A.
    Chakravarty S.
    Paikaray B.K.
    International Journal of Reasoning-based Intelligent Systems, 2023, 15 (01) : 15 - 21
  • [36] Frequency Based Named Entity Recognition System For Under Resource Language
    Debbarma, Abhijit
    Bhattacharya, Paritosh
    Purkayastha, B. S.
    2014 INTERNATIONAL CONFERENCE ON CONTROL, INSTRUMENTATION, COMMUNICATION AND COMPUTATIONAL TECHNOLOGIES (ICCICCT), 2014, : 847 - 849
  • [37] Mixup Based Cross-Consistency Training for Named Entity Recognition
    Youn, Geonsik
    Yoon, Bohan
    Ji, Seungbin
    Ko, Dahee
    Rhee, Jongtae
    APPLIED SCIENCES-BASEL, 2022, 12 (21):
  • [38] Pre-training and Evaluation of Numeracy-oriented Language Model
    Feng, Fuli
    Rui, Xilin
    Wang, Wenjie
    Cao, Yixin
    Chua, Tat-Seng
    ICAIF 2021: THE SECOND ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, 2021,
  • [39] A CRF based Machine Learning Approach for Biomedical Named Entity Recognition
    Kanimozhi, U.
    Manjula, D.
    2017 SECOND INTERNATIONAL CONFERENCE ON RECENT TRENDS AND CHALLENGES IN COMPUTATIONAL MODELS (ICRTCCM), 2017, : 335 - 342
  • [40] Korean named entity recognition based on language-specific features
    Chen, Yige
    Lim, KyungTae
    Park, Jungyeul
    NATURAL LANGUAGE ENGINEERING, 2024, 30 (03) : 625 - 649