A Mongolian-Chinese Neural Machine Translation Model Based on Soft Target Templates and Contextual Knowledge

被引:0
|
作者
Ren, Qing-Dao-Er-Ji [1 ]
Pang, Ziyu [1 ]
Lang, Jiajun [1 ]
机构
[1] Inner Mongolia Univ Technol, Sch Informat Engn, Hohhot 010051, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 21期
基金
中国国家自然科学基金;
关键词
neural machine translation; pre-training; contextual knowledge; soft target template;
D O I
10.3390/app132111845
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
In recent years, Mongolian-Chinese neural machine translation (MCNMT) technology has made substantial progress. However, the establishment of the Mongolian dataset requires a significant amount of financial and material investment, which has become a major obstacle to the performance of MCNMT. Pre-training and fine-tuning technology have also achieved great success in the field of natural language processing, but how to fully exploit the potential of pre-training language models (PLMs) in MCNMT has become an urgent problem to be solved. Therefore, this paper proposes a novel MCNMT model based on the soft target template and contextual knowledge. Firstly, to learn the grammatical structure of target sentences, a selection-based parsing tree is adopted to generate candidate templates that are used as soft target templates. The template information is merged with the encoder-decoder framework, fully utilizing the templates and source text information to guide the translation process. Secondly, the translation model learns the contextual knowledge of sentences from the BERT pre-training model through the dynamic fusion mechanism and knowledge extraction paradigm, so as to improve the model's utilization rate of language knowledge. Finally, the translation performance of the proposed model is further improved by integrating contextual knowledge and soft target templates by using a scaling factor. The effectiveness of the modified model is verified by a large number of data experiments, and the calculated BLEU (BiLingual Evaluation Understudy) value is increased by 4.032 points compared with the baseline MCNMT model of Transformers.
引用
收藏
页数:19
相关论文
共 50 条
  • [31] A Morpheme-Based Weighting for Chinese-Mongolian Statistical Machine Translation
    Yang, Zhenxin
    Li, Miao
    Chen, Lei
    Sun, Kai
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2016, E99D (11): : 2843 - 2846
  • [32] Incorporating Clause Alignment Knowledge into Chinese-English Neural Machine Translation
    Miao G.
    Liu M.
    Chen Y.
    Xu J.
    Zhang Y.
    Feng W.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2022, 58 (01): : 61 - 68
  • [33] Linguistic knowledge-based vocabularies for Neural Machine Translation
    Casas, Noe
    Costa-jussa, Marta R.
    Fonollosa, Jose A. R.
    Alonso, Juan A.
    Fanlo, Ramon
    NATURAL LANGUAGE ENGINEERING, 2021, 27 (04) : 485 - 506
  • [34] Transformer-Based Re-Ranking Model for Enhancing Contextual and Syntactic Translation in Low-Resource Neural Machine Translation
    Javed, Arifa
    Zan, Hongying
    Mamyrbayev, Orken
    Abdullah, Muhammad
    Ahmed, Kanwal
    Oralbekova, Dina
    Dinara, Kassymova
    Akhmediyarova, Ainur
    ELECTRONICS, 2025, 14 (02):
  • [35] Research on Machine Translation Model Based on Neural Network
    Han, Zhuoran
    Li, Shenghong
    COMMUNICATIONS, SIGNAL PROCESSING, AND SYSTEMS, CSPS 2018, VOL III: SYSTEMS, 2020, 517 : 244 - 251
  • [36] Optimization of Unsupervised Neural Machine Translation Based on Syntactic Knowledge Improvement
    Zhou, Aiping
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2023, 14 (11) : 90 - 99
  • [37] A Chinese-Malay Neural Machine Translation Model Based on CA-Transformer and Transfer Learning
    Zhan, Siqi
    Qin, Donghong
    Xu, Zhizhan
    Bao, Dongxue
    2022 IEEE THE 5TH INTERNATIONAL CONFERENCE ON BIG DATA AND ARTIFICIAL INTELLIGENCE (BDAI 2022), 2022, : 13 - 18
  • [38] Acquiring Knowledge from Pre-Trained Model to Neural Machine Translation
    Weng, Rongxiang
    Yu, Heng
    Huang, Shujian
    Cheng, Shanbo
    Luo, Weihua
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 9266 - 9273
  • [39] Post-editing neural machine translation versus phrase-based machine translation for English-Chinese
    Jia, Yanfang
    Carl, Michael
    Wang, Xiangling
    MACHINE TRANSLATION, 2019, 33 (1-2) : 9 - 29
  • [40] Incorporating Word Reordering Knowledge into Attention-based Neural Machine Translation
    Zhang, Jinchao
    Wang, Mingxuan
    Liu, Qun
    Zhou, Jie
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 1524 - 1534