Simultaneous neural machine translation with a reinforced attention mechanism

被引:6
|
作者
Lee, YoHan [1 ]
Shin, JongHun [1 ]
Kim, YoungKil [1 ]
机构
[1] Elect & Telecommun Res Inst, Language Intelligence Res Sect, Daejeon, South Korea
关键词
attention mechanism; neural network; reinforcement learning; simultaneous machine translation;
D O I
10.4218/etrij.2020-0358
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
To translate in real time, a simultaneous translation system should determine when to stop reading source tokens and generate target tokens corresponding to a partial source sentence read up to that point. However, conventional attention-based neural machine translation (NMT) models cannot produce translations with adequate latency in online scenarios because they wait until a source sentence is completed to compute alignment between the source and target tokens. To address this issue, we propose a reinforced learning (RL)-based attention mechanism, the reinforced attention mechanism, which allows a neural translation model to jointly train the stopping criterion and a partial translation model. The proposed attention mechanism comprises two modules, one to ensure translation quality and the other to address latency. Different from previous RL-based simultaneous translation systems, which learn the stopping criterion from a fixed NMT model, the modules can be trained jointly with a novel reward function. In our experiments, the proposed model has better translation quality and comparable latency compared to previous models.
引用
收藏
页码:775 / 786
页数:12
相关论文
共 50 条
  • [41] Look-Ahead Attention for Generation in Neural Machine Translation
    Zhou, Long
    Zhang, Jiajun
    Zong, Chengqing
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2017, 2018, 10619 : 211 - 223
  • [42] Neural Machine Translation With GRU-Gated Attention Model
    Zhang, Biao
    Xiong, Deyi
    Xie, Jun
    Su, Jinsong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (11) : 4688 - 4698
  • [43] English grammar correction based on attention mechanism machine translation
    Zhao, Kuaile
    International Journal of Computational Intelligence Studies, 2024, 13 (1-2) : 78 - 94
  • [44] Two-Way Neural Network Chinese-English Machine Translation Model Fused with Attention Mechanism
    Liang, Jing
    Du, Minghui
    SCIENTIFIC PROGRAMMING, 2022, 2022
  • [45] RETRACTED: Research on the Construction of a Bidirectional Neural Network Machine Translation Model Fused with Attention Mechanism (Retracted Article)
    Zuo, Guangming
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2022, 2022
  • [46] Machine Translation for Indian Languages Utilizing Recurrent Neural Networks and Attention
    Sharma, Sonali
    Diwakar, Manoj
    DISTRIBUTED COMPUTING AND OPTIMIZATION TECHNIQUES, ICDCOT 2021, 2022, 903 : 593 - 602
  • [47] Hybrid Attention for Chinese Character-Level Neural Machine Translation
    Wang, Feng
    Chen, Wei
    Yang, Zhen
    Xu, Shuang
    Xu, Bo
    NEUROCOMPUTING, 2019, 358 : 44 - 52
  • [48] Neural Machine Translation Models with Attention-Based Dropout Layer
    Israr, Huma
    Khan, Safdar Abbas
    Tahir, Muhammad Ali
    Shahzad, Muhammad Khuram
    Ahmad, Muneer
    Zain, Jasni Mohamad
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 75 (02): : 2981 - 3009
  • [49] Neural Machine Translation with Attention Based on a New Syntactic Branch Distance
    Peng, Ru
    Chen, Zhitao
    Hao, Tianyong
    Fang, Yi
    MACHINE TRANSLATION, CCMT 2019, 2019, 1104 : 47 - 57
  • [50] Multi-Granularity Self-Attention for Neural Machine Translation
    Hao, Jie
    Wang, Xing
    Shi, Shuming
    Zhang, Jinfeng
    Tu, Zhaopeng
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 887 - 897