Efficient Wait-k Models for Simultaneous Machine Translation

被引:23
作者
Elbayad, Maha [1 ]
Besacier, Laurent [1 ]
Verbeek, Jakob [2 ]
机构
[1] Univ Grenoble Alpes, CNRS, Grenoble INP, Inria,LIG,LJK, F-38000 Grenoble, France
[2] Facebook AI Res, Menlo Pk, CA USA
来源
INTERSPEECH 2020 | 2020年
关键词
D O I
10.21437/Interspeech.2020-1241
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Simultaneous machine translation consists in starting output generation before the entire input sequence is available. Wait-k decoders offer a simple but efficient approach for this problem. They first read k source tokens, after which they alternate between producing a target token and reading another source token. We investigate the behavior of wait-k decoding in low resource settings for spoken corpora using IWSLT datasets. We improve training of these models using unidirectional encoders, and training across multiple values of k. Experiments with Transformer and 2D-convolutional architectures show that our wait-k models generalize well across a wide range of latency levels. We also show that the 2D-convolution architecture is competitive with Transformers for simultaneous translation of spoken language.
引用
收藏
页码:1461 / 1465
页数:5
相关论文
共 30 条
[1]  
Alinejad A., 2018, P EMNLP
[2]  
[Anonymous], 2017, P ICASSP
[3]  
Arivazhagan N., 2020, P IWSLT
[4]  
Arivazhagan N, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P1313
[5]  
Cettolo M., 2014, P IWSLT
[6]  
Chen M., 2018, P ACL
[7]  
Chiu C.-C., 2018, 6 INT C LEARN REPR I
[8]  
Cho Kyunghyun, 2016, Can neural machine translation do simultaneous translation?
[9]  
Dalvi F, 2018, P 2018 C N AM CHAPT, V2, P493, DOI DOI 10.18653/V1/N18-2079
[10]  
Edunov S., 2018, P NAACL HLT