Alleviating repetitive tokens in non-autoregressive machine translation with unlikelihood training

被引:0
|
作者
Shuheng Wang
Shumin Shi
Heyan Huang
机构
[1] Nanyang Institute of Technology,School of Computer and Software
[2] Beijing Institute of Technology,School of Computer Science and Technology
来源
Soft Computing | 2024年 / 28卷
关键词
Machine translation; Non-autoregressive; Repetitive tokens; Unlikelihood training;
D O I
暂无
中图分类号
学科分类号
摘要
In recent years, significant progress has been made in the field of non-autoregressive machine translations. However, the accuracy of non-autoregressive models still lags behind their autoregressive counterparts. This discrepancy can be attributed to the abundance of repetitive tokens in the target sequences generated by non-autoregressive models. In this study, we delve into this phenomenon and propose a novel approach to train a non-autoregressive model using unlikelihood loss. We evaluate our method on three widely used benchmark tasks. The experimental results demonstrating that our proposed approach significantly reduces the number of repetitive tokens while improving the overall performance of non-autoregressive machine translations. Compared to the baseline model ”Mask-Predict”, the average number of repetitions on IWSLT 14 DE→\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\rightarrow $$\end{document}EN valid set is reduced from 0.48 to 0.17, resulting in a remarkable 62% decrease.
引用
收藏
页码:4681 / 4688
页数:7
相关论文
共 50 条
  • [31] Non-Autoregressive Neural Machine Translation with Enhanced Decoder Input
    Guo, Junliang
    Tan, Xu
    He, Di
    Qin, Tao
    Xu, Linli
    Liu, Tie-Yan
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 3723 - 3730
  • [32] Progressive Multi-Granularity Training for Non-Autoregressive Translation
    Ding, Liang
    Wang, Longyue
    Liu, Xuebo
    Wong, Derek F.
    Tao, Dacheng
    Tu, Zhaopeng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 2797 - 2803
  • [33] Retrieving Sequential Information for Non-Autoregressive Neural Machine Translation
    Shao, Chenze
    Feng, Yang
    Zhang, Jinchao
    Meng, Fandong
    Chen, Xilin
    Zhou, Jie
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3013 - 3024
  • [34] Correcting translation for non-autoregressive transformer
    Wang, Shuheng
    Huang, Heyan
    Shi, Shumin
    Li, Dongbai
    Guo, Dongen
    APPLIED SOFT COMPUTING, 2025, 168
  • [35] Order-Agnostic Cross Entropy for Non-Autoregressive Machine Translation
    Du, Cunxiao
    Tu, Zhaopeng
    Jiang, Jing
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [36] A Study of Syntactic Multi-Modality in Non-Autoregressive Machine Translation
    Zhang, Kexun
    Wang, Rui
    Tan, Xu
    Guo, Junliang
    Ren, Yi
    Qin, Tao
    Liu, Tie-Yan
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 1747 - 1757
  • [37] Revisiting Non-Autoregressive Translation at Scale
    Wang, Zhihao
    Wang, Longyue
    Su, Jinsong
    Yao, Junfeng
    Tu, Zhaopeng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 12051 - 12065
  • [38] Improving Non-autoregressive Machine Translation with Error Exposure and Consistency Regularization
    Chen, Xinran
    Duan, Sufeng
    Liu, Gongshen
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT III, NLPCC 2024, 2025, 15361 : 240 - 252
  • [39] Iterative Refinement in the Continuous Space for Non-Autoregressive Neural Machine Translation
    Lee, Jason
    Shu, Raphael
    Cho, Kyunghyun
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1006 - 1015
  • [40] Non-autoregressive Machine Translation with Probabilistic Context-free Grammar
    Gui, Shangtong
    Shao, Chenze
    Ma, Zhengrui
    Zhang, Xishan
    Chen, Yunji
    Feng, Yang
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,