Pre-training neural machine translation with alignment information via optimal transport

被引:0
|
作者
Su, Xueping [1 ]
Zhao, Xingkai [1 ]
Ren, Jie [1 ]
Li, Yunhong [1 ]
Raetsch, Matthias [2 ]
机构
[1] Xian Polytech Univ, Sch Elect & Informat, Xian, Peoples R China
[2] Reutlingen Univ, Dept Engn, Interact & Mobile Robot & Artificial Intelligence, Reutlingen, Germany
基金
中国国家自然科学基金;
关键词
Optimal Transport; Alignment Information; Pre-training; Neural Machine Translation;
D O I
10.1007/s11042-023-17479-z
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the rapid development of globalization, the demand for translation between different languages is also increasing. Although pre-training has achieved excellent results in neural machine translation, the existing neural machine translation has almost no high-quality suitable for specific fields. Alignment information, so this paper proposes a pre-training neural machine translation with alignment information via optimal transport. First, this paper narrows the representation gap between different languages by using OTAP to generate domain-specific data for information alignment, and learns richer semantic information. Secondly, this paper proposes a lightweight model DR-Reformer, which uses Reformer as the backbone network, adds Dropout layers and Reduction layers, reduces model parameters without losing accuracy, and improves computational efficiency. Experiments on the Chinese and English datasets of AI Challenger 2018 and WMT-17 show that the proposed algorithm has better performance than existing algorithms.
引用
收藏
页码:48377 / 48397
页数:21
相关论文
共 50 条
  • [11] Pre-training on dynamic graph neural networks
    Chen, Ke-Jia
    Zhang, Jiajun
    Jiang, Linpu
    Wang, Yunyun
    Dai, Yuxuan
    NEUROCOMPUTING, 2022, 500 : 679 - 687
  • [12] Product-oriented Machine Translation with Cross-modal Cross-lingual Pre-training
    Song, Yuqing
    Chen, Shizhe
    Jin, Qin
    Luo, Wei
    Xie, Jun
    Huang, Fei
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 2843 - 2852
  • [13] PHGNN: Pre-Training Heterogeneous Graph Neural Networks
    Li, Xin
    Wei, Hao
    Ding, Yu
    IEEE ACCESS, 2024, 12 : 135411 - 135418
  • [14] Webformer: Pre-training with Web Pages for Information Retrieval
    Guo, Yu
    Ma, Zhengyi
    Mao, Jiaxin
    Qian, Hongjin
    Zhang, Xinyu
    Jiang, Hao
    Cao, Zhao
    Dou, Zhicheng
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 1502 - 1512
  • [15] Improving neural machine translation with sentence alignment learning
    Shi, Xuewen
    Huang, Heyan
    Jian, Ping
    Tang, Yi-Kun
    NEUROCOMPUTING, 2021, 420 : 15 - 26
  • [16] PF-HIN:Pre-Training for Heterogeneous Information Networks
    Fang, Yang
    Zhao, Xiang
    Chen, Yifan
    Xiao, Weidong
    de Rijke, Maarten
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (08) : 8372 - 8385
  • [17] Pre-Training of an Artificial Neural Network for Software Fault Prediction
    Owhadi-Kareshk, Moein
    Sedaghat, Yasser
    Akbarzadeh-T, Mohammad-R
    PROCEEDINGS OF THE 2017 7TH INTERNATIONAL CONFERENCE ON COMPUTER AND KNOWLEDGE ENGINEERING (ICCKE), 2017, : 223 - 228
  • [18] Pre-training Graph Neural Network for Cross Domain Recommendation
    Wang, Chen
    Liang, Yueqing
    Liu, Zhiwei
    Zhang, Tao
    Yu, Philip S.
    2021 IEEE THIRD INTERNATIONAL CONFERENCE ON COGNITIVE MACHINE INTELLIGENCE (COGMI 2021), 2021, : 140 - 145
  • [19] Vision-language pre-training via modal interaction
    Cheng, Hang
    Ye, Hehui
    Zhou, Xiaofei
    Liu, Ximeng
    Chen, Fei
    Wang, Meiqing
    PATTERN RECOGNITION, 2024, 156
  • [20] Generative adversarial training for neural machine translation
    Yang, Zhen
    Chen, Wei
    Wang, Feng
    Xu, Bo
    NEUROCOMPUTING, 2018, 321 : 146 - 155