LKMT: Linguistics Knowledge-Driven Multi-Task Neural Machine Translation for Urdu and English

被引:0
|
作者
Hassan, Muhammad Naeem Ul [1 ,2 ]
Yu, Zhengtao [1 ,2 ]
Wang, Jian [1 ,2 ]
Li, Ying [1 ,2 ]
Gao, Shengxiang [1 ,2 ]
Yang, Shuwan [1 ,2 ]
Mao, Cunli [1 ,2 ]
机构
[1] Kunming Univ Sci & Technol, Fac Informat Engn & Automat, Kunming 650500, Peoples R China
[2] Kunming Univ Sci & Technol, Yunnan Key Lab Artificial Intelligence, Kunming 650500, Peoples R China
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2024年 / 81卷 / 01期
基金
中国国家自然科学基金;
关键词
Urdu NMT (neural machine translation); Urdu natural language processing; Urdu Linguistic features; low resources language; linguistic features pretrain model;
D O I
10.32604/cmc.2024.054673
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Thanks to the strong representation capability of pre-trained language models, supervised machine translation models have achieved outstanding performance. However, the performances of these models drop sharply when the scale of the parallel training corpus is limited. Considering the pre-trained language model has a strong ability for monolingual representation, it is the key challenge for machine translation to construct the in-depth relationship between the source and target language by injecting the lexical and syntactic information into pre-trained language models. To alleviate the dependence on the parallel corpus, we propose a Linguistics Knowledge-Driven Multi- Task (LKMT) approach to inject part-of-speech and syntactic knowledge into pre-trained models, thus enhancing the machine translation performance. On the one hand, we integrate part-of-speech and dependency labels into the embedding layer and exploit large-scale monolingual corpus to update all parameters of pre-trained language models, thus ensuring the updated language model contains potential lexical and syntactic information. On the other hand, we leverage an extra self-attention layer to explicitly inject linguistic knowledge into the pre-trained language model-enhanced machine translation model. Experiments on the benchmark dataset show that our proposed LKMT approach improves the Urdu-English translation accuracy by 1.97 points and the English-Urdu translation accuracy by 2.42 points, highlighting the effectiveness of our LKMT framework. Detailed ablation experiments confirm the positive impact of part-of-speech and dependency parsing on machine translation.
引用
收藏
页码:951 / 969
页数:19
相关论文
共 49 条
  • [31] MT2: Towards a Multi-Task Machine Translation Model with Translation-Specific In-Context Learning
    Li, Chunyou
    Liu, Mingtong
    Zhang, Hongxiao
    Chen, Yufeng
    Xu, Jinan
    Zhou, Ming
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 8616 - 8627
  • [32] TransMKR: Translation-based knowledge graph enhanced multi-task point-of-interest recommendation
    Hu, Bojing
    Ye, Yaqin
    Zhong, Yingqiang
    Pan, Jiao
    Hu, Maosheng
    NEUROCOMPUTING, 2022, 474 : 107 - 114
  • [33] Multi-Source Transformer for Kazakh-Russian-English Neural Machine Translation
    Littell, Patrick
    Lo, Chi-kiu
    Larkin, Samuel
    Stewart, Darlene
    FOURTH CONFERENCE ON MACHINE TRANSLATION (WMT 2019), 2019, : 267 - 274
  • [34] Test-Driven Multi-Task Learning with Functionally Equivalent Code Transformation for Neural Code Generation
    Wang, Xin
    Liu, Xiao
    Zhou, Pingyi
    Liu, Qixia
    Liu, Jin
    Wu, Hao
    Cui, Xiaohui
    PROCEEDINGS OF THE 37TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE 2022, 2022,
  • [35] MACHINE LEARNING USING A MULTI-TASK CONVOLUTIONAL NEURAL NETWORKS CAN ACCURATELY ASSESS ROBOTIC SKILLS
    Gahan, Jeffrey
    Steinberg, Ryan
    Garbens, Alaina
    Qu, Xingming
    Larson, Eric
    JOURNAL OF UROLOGY, 2020, 203 : E505 - E505
  • [36] Multi-granularity Knowledge Sharing in Low-resource Neural Machine Translation
    Mi, Chenggang
    Xie, Shaoliang
    Fan, Yi
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2024, 23 (02)
  • [37] Building a Multi-Domain Neural Machine Translation Model Using Knowledge Distillation
    Mghabbar, Idriss
    Ratnamogan, Pirashanth
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 2116 - 2123
  • [38] Multi-Task Learning in Deep Neural Networks for Mandarin-English Code-Mixing Speech Recognition
    Chen, Mengzhe
    Pan, Jielin
    Zhao, Qingwei
    Yan, Yonghong
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2016, E99D (10): : 2554 - 2557
  • [39] A multi-stack RNN-based neural machine translation model for English to Pakistan sign language translation
    Uzma Farooq
    Mohd Shafry Mohd Rahim
    Adnan Abid
    Neural Computing and Applications, 2023, 35 : 13225 - 13238
  • [40] A multi-stack RNN-based neural machine translation model for English to Pakistan sign language translation
    Farooq, Uzma
    Rahim, Mohd Shafry Mohd
    Abid, Adnan
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (18): : 13225 - 13238