Neural Machine Translation With Sentence-Level Topic Context

被引:33
作者
Chen, Kehai [1 ]
Wang, Rui [2 ]
Utiyama, Masao [2 ]
Sumita, Eiichiro [2 ]
Zhao, Tiejun [1 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Technol, Machine Intelligence & Translat Lab, Harbin 150001, Heilongjiang, Peoples R China
[2] Natl Inst Informat & Commun Technol, Adv Speech Translat Res & Dev Promot Ctr, Kyoto 6190289, Japan
关键词
Sentence-level Context; Latent Topic Representation; Convolutional Neural Network; Neural Machine Translation;
D O I
10.1109/TASLP.2019.2937190
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Traditional neural machine translation (NMT) methods use theword-level context to predict target language translation while neglecting the sentence-level context, which has been shown to be beneficial for translation prediction in statistical machine translation. This paper represents the sentence-level context as latent topic representations by using a convolution neural network, and designs a topic attention to integrate source sentence-level topic context information into both attention-based and Transformerbased NMT. In particular, our method can improve the performance of NMT by modeling source topics and translations jointly. Experiments on the large-scale LDC Chinese-to-English translation tasks and WMT'14 English-to-German translation tasks show that the proposed approach can achieve significant improvements compared with baseline systems.
引用
收藏
页码:1970 / 1984
页数:15
相关论文
共 60 条
[1]  
[Anonymous], 2017, P ICLR
[2]  
[Anonymous], 2018, P 27 INT C COMP LING
[3]  
[Anonymous], 2017, P SOFTW DEM 15 C EUR, DOI DOI 10.18653/V1
[4]  
[Anonymous], 2016, C ASS MACHINE TRANSL
[5]  
[Anonymous], 2005, P ACL
[6]  
[Anonymous], 2016, P NAACL HLT
[7]  
[Anonymous], 2018, P 27 INT C COMP LING
[8]  
[Anonymous], 2017, EMNLP
[9]  
Bahdanau Dzmitry, 2015, 3 INT C LEARNING REP
[10]   Doubly-Attentive Decoder for Multi-modal Neural Machine Translation [J].
Calixto, Iacer ;
Liu, Qun ;
Campbell, Nick .
PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, :1913-1924