LKMT: Linguistics Knowledge-Driven Multi-Task Neural Machine Translation for Urdu and English

被引:0
|
作者
Hassan, Muhammad Naeem Ul [1 ,2 ]
Yu, Zhengtao [1 ,2 ]
Wang, Jian [1 ,2 ]
Li, Ying [1 ,2 ]
Gao, Shengxiang [1 ,2 ]
Yang, Shuwan [1 ,2 ]
Mao, Cunli [1 ,2 ]
机构
[1] Kunming Univ Sci & Technol, Fac Informat Engn & Automat, Kunming 650500, Peoples R China
[2] Kunming Univ Sci & Technol, Yunnan Key Lab Artificial Intelligence, Kunming 650500, Peoples R China
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2024年 / 81卷 / 01期
基金
中国国家自然科学基金;
关键词
Urdu NMT (neural machine translation); Urdu natural language processing; Urdu Linguistic features; low resources language; linguistic features pretrain model;
D O I
10.32604/cmc.2024.054673
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Thanks to the strong representation capability of pre-trained language models, supervised machine translation models have achieved outstanding performance. However, the performances of these models drop sharply when the scale of the parallel training corpus is limited. Considering the pre-trained language model has a strong ability for monolingual representation, it is the key challenge for machine translation to construct the in-depth relationship between the source and target language by injecting the lexical and syntactic information into pre-trained language models. To alleviate the dependence on the parallel corpus, we propose a Linguistics Knowledge-Driven Multi- Task (LKMT) approach to inject part-of-speech and syntactic knowledge into pre-trained models, thus enhancing the machine translation performance. On the one hand, we integrate part-of-speech and dependency labels into the embedding layer and exploit large-scale monolingual corpus to update all parameters of pre-trained language models, thus ensuring the updated language model contains potential lexical and syntactic information. On the other hand, we leverage an extra self-attention layer to explicitly inject linguistic knowledge into the pre-trained language model-enhanced machine translation model. Experiments on the benchmark dataset show that our proposed LKMT approach improves the Urdu-English translation accuracy by 1.97 points and the English-Urdu translation accuracy by 2.42 points, highlighting the effectiveness of our LKMT framework. Detailed ablation experiments confirm the positive impact of part-of-speech and dependency parsing on machine translation.
引用
收藏
页码:951 / 969
页数:19
相关论文
共 49 条
  • [21] Multi-Task Learning with Shared Encoder for Non-Autoregressive Machine Translation
    Hao, Yongchang
    He, Shilin
    Jiao, Wenxiang
    Tu, Zhaopeng
    Lyu, Michael R.
    Wang, Xing
    2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 3989 - 3996
  • [22] Evolutionary Multi-task Learning for Modular Knowledge Representation in Neural Networks
    Chandra, Rohitash
    Gupta, Abhishek
    Ong, Yew-Soon
    Goh, Chi-Keong
    NEURAL PROCESSING LETTERS, 2018, 47 (03) : 993 - 1009
  • [23] Evolutionary Multi-task Learning for Modular Knowledge Representation in Neural Networks
    Rohitash Chandra
    Abhishek Gupta
    Yew-Soon Ong
    Chi-Keong Goh
    Neural Processing Letters, 2018, 47 : 993 - 1009
  • [24] Knowledge Adaption for Demand Prediction based on Multi-task Memory Neural Network
    Li, Can
    Bai, Lei
    Liu, Wei
    Yao, Lina
    Waller, S. Travis
    CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 715 - 724
  • [25] Multi-task convolutional deep neural network for recommendation based on knowledge graphs
    Jiang, Mingyang
    Li, Man
    Cao, Wenming
    Yang, Mingming
    Zhou, Luxin
    NEUROCOMPUTING, 2025, 619
  • [26] Incorporating Clause Alignment Knowledge into Chinese-English Neural Machine Translation
    Miao G.
    Liu M.
    Chen Y.
    Xu J.
    Zhang Y.
    Feng W.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2022, 58 (01): : 61 - 68
  • [27] Improving Neural Machine Translation by Multi-Knowledge Integration with Prompting
    Wang, Ke
    Xie, Jun
    Zhang, Yuqi
    Zhao, Yu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 5000 - 5010
  • [28] Attribute Knowledge Integration for Speech Recognition Based on Multi-task Learning Neural Networks
    Zheng, Hao
    Yang, Zhanlei
    Qiao, Liwei
    Li, Jianping
    Liu, Wenju
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 543 - 547
  • [29] Multi-Task Neural Network for Non-discrete Attribute Prediction in Knowledge Graphs
    Tay, Yi
    Luu Anh Tuan
    Phan, Minh C.
    Hui, Siu Cheung
    CIKM'17: PROCEEDINGS OF THE 2017 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2017, : 1029 - 1038
  • [30] A Sequential Multi-task Learning Neural Network with Metric-Based Knowledge Transfer
    Yue, Simeng
    Ozawa, Seiichi
    2012 11TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2012), VOL 1, 2012, : 671 - 674