English-Assamese neural machine translation using prior alignment and pre-trained language model

被引:9
作者
Laskar, Sahinur Rahman [1 ]
Paul, Bishwaraj [1 ]
Dadure, Pankaj [2 ]
Manna, Riyanka [3 ]
Pakray, Partha [1 ]
Bandyopadhyay, Sivaji [1 ]
机构
[1] Natl Inst Technol, Dept Comp Sci & Engn, Silchar 788010, Assam, India
[2] Univ Petr & Energy Studies, Sch Comp Sci, Dehra Dun 248007, Uttarakhand, India
[3] Jadavpur Univ, Dept Comp Sci & Engn, Kolkata 700032, W Bengal, India
关键词
Low-resource; NMT; English-Assamese; Alignment; Language model;
D O I
10.1016/j.csl.2023.101524
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In a multilingual country like India, automatic natural language translation plays a key role in building a community with different linguistic people. Many researchers have explored and improved the translation process for high-resource languages such as English, German, etc., and achieved state-of-the-art results. However, the unavailability of adequate data is the prime obstacle to automatic natural language translation of low-resource north-eastern Indian languages such as Mizo, Khasi, and Assamese. Though the recent past has witnessed a deluge in several automatic natural language translation systems for low-resource languages, the low values of their evaluation measures indicate the scope for improvement. In the recent past, the neural machine translation approach has significantly improved translation quality, and the credit goes to the availability of a huge amount of data. Subsequently, the neural machine translation approach for low-resource language is underrepresented due to the unavailability of adequate data. In this work, we have considered a low-resource English-Assamese pair using the transformer-based neural machine translation, which leverages the use of prior alignment and a pre-trained language model. To extract alignment information from the source-target sentences, we have used the pre-trained multilingual contextual embeddings-based alignment technique. Also, the transformer-based language model is built using monolingual target sentences. With the use of both prior alignment and a pre-trained language model, the transformer-based neural machine translation model shows improvement, and we have achieved state-of-the-art results for the English-to-Assamese and Assamese-to-English translation, respectively.
引用
收藏
页数:16
相关论文
共 32 条
  • [21] Conversational Model Using Neural Machine Translation
    Ramu, Agusthiyar
    Gokul, R.
    Sumesh, Nihal
    [J]. BIOSCIENCE BIOTECHNOLOGY RESEARCH COMMUNICATIONS, 2020, 13 (04): : 159 - 162
  • [22] Text based personality prediction from multiple social media data sources using pre-trained language model and model averaging
    Hans Christian
    Derwin Suhartono
    Andry Chowanda
    Kamal Z. Zamli
    [J]. Journal of Big Data, 8
  • [23] An empirical analysis on statistical and neural machine translation system for English to Mizo language
    Devi C.S.
    Purkayastha B.S.
    [J]. International Journal of Information Technology, 2023, 15 (8) : 4021 - 4028
  • [24] Using Pre-trained Language Models to Resolve Textual and Semantic Merge Conflicts (Experience Paper)
    Zhang, Jialu
    Mytkowicz, Todd
    Kaufman, Mike
    Piskac, Ruzica
    Lahiri, Shuvendu K.
    [J]. PROCEEDINGS OF THE 31ST ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2022, 2022, : 77 - 88
  • [25] Text based personality prediction from multiple social media data sources using pre-trained language model and model averaging
    Christian, Hans
    Suhartono, Derwin
    Chowanda, Andry
    Zamli, Kamal Z.
    [J]. JOURNAL OF BIG DATA, 2021, 8 (01)
  • [26] EMS-BERT: A Pre-Trained Language Representation Model for the Emergency Medical Services (EMS) Domain
    Rahman, M. Arif
    Preum, Sarah Masud
    Williams, Ronald D.
    Alemzadeh, Homa
    Stankovic, John
    [J]. 2023 IEEE/ACM CONFERENCE ON CONNECTED HEALTH: APPLICATIONS, SYSTEMS AND ENGINEERING TECHNOLOGIES, CHASE, 2023, : 34 - 43
  • [27] Low-Resource Neural Machine Translation Using XLNet Pre-training Model
    Wu, Nier
    Hou, Hongxu
    Guo, Ziyue
    Zheng, Wei
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2021, PT V, 2021, 12895 : 503 - 514
  • [28] COVID-HateBERT: a Pre-trained Language Model for COVID-19 related Hate Speech Detection
    Li, Mingqi
    Liao, Song
    Okpala, Ebuka
    Tong, Max
    Costello, Matthew
    Cheng, Long
    Hu, Hongxin
    Luo, Feng
    [J]. 20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 233 - 238
  • [29] Decoding Optimization for Chinese-English Machine Translation via a Dependent Syntax Language Model
    Liu, Ying
    Yu, Zhengtao
    Zhang, Tao
    Zhao, Xing
    [J]. ARTIFICIAL INTELLIGENCE AND COMPUTATIONAL INTELLIGENCE, PT III, 2011, 7004 : 143 - 150
  • [30] Language Model Pre-training Method in Machine Translation Based on Named Entity Recognition
    Li, Zhen
    Qu, Dan
    Xie, Chaojie
    Zhang, Wenlin
    Li, Yanxia
    [J]. INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, 2020, 29 (7-8)