Linguistically Driven Multi-Task Pre-Training for Low-Resource Neural Machine Translation

被引:7
|
作者
Mao, Zhuoyuan [1 ]
Chu, Chenhui [1 ]
Kurohashi, Sadao [1 ]
机构
[1] Kyoto Univ, Grad Sch Informat, Kyoto, Japan
关键词
Low-resource neural machine translation; pre-training; linguistically-driven;
D O I
10.1145/3491065
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the present study, we propose novel sequence-to-sequence pre-training objectives for low-resource machine translation (NMT): Japanese-specific sequence to sequence (JASS) for language pairs involving Japanese as the source or target language, and English-specific sequence to sequence (ENSS) for language pairs involving English. JASS focuses on masking and reordering Japanese linguistic units known as bunsetsu, whereas ENSS is proposed based on phrase structure masking and reordering tasks. Experiments on ASPEC Japanese-English & Japanese-Chinese, Wikipedia Japanese-Chinese, News English-Korean corpora demonstrate that JASS and ENSS outperform MASS and other existing language-agnostic pre-training methods by up to +2.9 BLEU points for the Japanese-English tasks, up to +7.0 BLEU points for the Japanese-Chinese tasks and up to +1.3 BLEU points for English-Korean tasks. Empirical analysis, which focuses on the relationship between individual parts in JASS and ENSS, reveals the complementary nature of the subtasks of JASS and ENSS. Adequacy evaluation using LASER, human evaluation, and case studies reveals that our proposed methods significantly outperform pre-training methods without injected linguistic knowledge and they have a larger positive impact on the adequacy as compared to the fluency.
引用
收藏
页数:29
相关论文
共 25 条
  • [21] Text-to-Speech for Low-Resource Agglutinative Language With Morphology-Aware Language Model Pre-Training
    Liu, Rui
    Hu, Yifan
    Zuo, Haolin
    Luo, Zhaojie
    Wang, Longbiao
    Gao, Guanglai
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 1075 - 1087
  • [22] Cross-dataset transfer learning for motor imagery signal classification via multi-task learning and pre-training
    Xie, Yuting
    Wang, Kun
    Meng, Jiayuan
    Yue, Jin
    Meng, Lin
    Yi, Weibo
    Jung, Tzyy-Ping
    Xu, Minpeng
    Ming, Dong
    JOURNAL OF NEURAL ENGINEERING, 2023, 20 (05)
  • [23] Distilling BERT knowledge into Seq2Seq with regularized Mixup for low-resource neural machine translation
    Zhang, Guanghua
    Liu, Hua
    Guo, Junjun
    Guo, Tianyu
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 259
  • [24] Product-oriented Machine Translation with Cross-modal Cross-lingual Pre-training
    Song, Yuqing
    Chen, Shizhe
    Jin, Qin
    Luo, Wei
    Xie, Jun
    Huang, Fei
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 2843 - 2852
  • [25] Multi-label rhinitis prediction using ensemble neural network chain with pre-training
    Yang, Jingdong
    Zhang, Meng
    Liu, Peng
    Yu, Shaoqing
    APPLIED SOFT COMPUTING, 2022, 122