Rephrasing the Reference for Non-autoregressive Machine Translation

被引:0
|
作者
Shao, Chenze [1 ,2 ]
Zhang, Jinchao [3 ]
Zhou, Jie [3 ]
Feng, Yang [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, Key Lab Intelligent Informat Proc, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Tencent Inc, Pattern Recognit Ctr, WeChat AI, Shenzhen, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Non-autoregressive neural machine translation (NAT) models suffer from the multi-modality problem that there may exist multiple possible translations of a source sentence, so the reference sentence may be inappropriate for the training when the NAT output is closer to other translations. In response to this problem, we introduce a rephraser to provide a better training target for NAT by rephrasing the reference sentence according to the NAT output. As we train NAT based on the rephraser output rather than the reference sentence, the rephraser output should fit well with the NAT output and not deviate too far from the reference, which can be quantified as reward functions and optimized by reinforcement learning. Experiments on major WMT benchmarks and NAT baselines show that our approach consistently improves the translation quality of NAT. Specifically, our best variant achieves comparable performance to the autoregressive Transformer, while being 14.7 times more efficient in inference.
引用
收藏
页码:13538 / 13546
页数:9
相关论文
共 50 条
  • [31] Correcting translation for non-autoregressive transformer
    Wang, Shuheng
    Huang, Heyan
    Shi, Shumin
    Li, Dongbai
    Guo, Dongen
    APPLIED SOFT COMPUTING, 2025, 168
  • [32] Order-Agnostic Cross Entropy for Non-Autoregressive Machine Translation
    Du, Cunxiao
    Tu, Zhaopeng
    Jiang, Jing
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [33] A Study of Syntactic Multi-Modality in Non-Autoregressive Machine Translation
    Zhang, Kexun
    Wang, Rui
    Tan, Xu
    Guo, Junliang
    Ren, Yi
    Qin, Tao
    Liu, Tie-Yan
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 1747 - 1757
  • [34] Alleviating repetitive tokens in non-autoregressive machine translation with unlikelihood training
    Wang, Shuheng
    Shi, Shumin
    Huang, Heyan
    SOFT COMPUTING, 2024, 28 (5) : 4681 - 4688
  • [35] Alleviating repetitive tokens in non-autoregressive machine translation with unlikelihood training
    Shuheng Wang
    Shumin Shi
    Heyan Huang
    Soft Computing, 2024, 28 : 4681 - 4688
  • [36] Revisiting Non-Autoregressive Translation at Scale
    Wang, Zhihao
    Wang, Longyue
    Su, Jinsong
    Yao, Junfeng
    Tu, Zhaopeng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 12051 - 12065
  • [37] Sequence-Level Training for Non-Autoregressive Neural Machine Translation
    Shao, Chenze
    Feng, Yang
    Zhang, Jinchao
    Meng, Fandong
    Zhou, Jie
    COMPUTATIONAL LINGUISTICS, 2021, 47 (04) : 891 - 925
  • [38] Improving Non-autoregressive Machine Translation with Error Exposure and Consistency Regularization
    Chen, Xinran
    Duan, Sufeng
    Liu, Gongshen
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT III, NLPCC 2024, 2025, 15361 : 240 - 252
  • [39] Iterative Refinement in the Continuous Space for Non-Autoregressive Neural Machine Translation
    Lee, Jason
    Shu, Raphael
    Cho, Kyunghyun
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1006 - 1015
  • [40] Non-autoregressive Machine Translation with Probabilistic Context-free Grammar
    Gui, Shangtong
    Shao, Chenze
    Ma, Zhengrui
    Zhang, Xishan
    Chen, Yunji
    Feng, Yang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,