English-Assamese neural machine translation using prior alignment and pre-trained language model

被引:9
|
作者
Laskar, Sahinur Rahman [1 ]
Paul, Bishwaraj [1 ]
Dadure, Pankaj [2 ]
Manna, Riyanka [3 ]
Pakray, Partha [1 ]
Bandyopadhyay, Sivaji [1 ]
机构
[1] Natl Inst Technol, Dept Comp Sci & Engn, Silchar 788010, Assam, India
[2] Univ Petr & Energy Studies, Sch Comp Sci, Dehra Dun 248007, Uttarakhand, India
[3] Jadavpur Univ, Dept Comp Sci & Engn, Kolkata 700032, W Bengal, India
关键词
Low-resource; NMT; English-Assamese; Alignment; Language model;
D O I
10.1016/j.csl.2023.101524
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In a multilingual country like India, automatic natural language translation plays a key role in building a community with different linguistic people. Many researchers have explored and improved the translation process for high-resource languages such as English, German, etc., and achieved state-of-the-art results. However, the unavailability of adequate data is the prime obstacle to automatic natural language translation of low-resource north-eastern Indian languages such as Mizo, Khasi, and Assamese. Though the recent past has witnessed a deluge in several automatic natural language translation systems for low-resource languages, the low values of their evaluation measures indicate the scope for improvement. In the recent past, the neural machine translation approach has significantly improved translation quality, and the credit goes to the availability of a huge amount of data. Subsequently, the neural machine translation approach for low-resource language is underrepresented due to the unavailability of adequate data. In this work, we have considered a low-resource English-Assamese pair using the transformer-based neural machine translation, which leverages the use of prior alignment and a pre-trained language model. To extract alignment information from the source-target sentences, we have used the pre-trained multilingual contextual embeddings-based alignment technique. Also, the transformer-based language model is built using monolingual target sentences. With the use of both prior alignment and a pre-trained language model, the transformer-based neural machine translation model shows improvement, and we have achieved state-of-the-art results for the English-to-Assamese and Assamese-to-English translation, respectively.
引用
收藏
页数:16
相关论文
共 32 条
  • [1] Multimodal Neural Machine Translation for English-Assamese Pair
    Laskar, Sahinur Rahman
    Paul, Bishwaraj
    Paudwal, Siddharth
    Gautam, Pranjit
    Biswas, Nirmita
    Pakray, Partha
    2021 INTERNATIONAL CONFERENCE ON COMPUTATIONAL PERFORMANCE EVALUATION (COMPE-2021), 2021, : 387 - 392
  • [2] A Domain Specific Parallel Corpus and Enhanced English-Assamese Neural Machine Translation
    Laskar, Sahinur Rahman
    Manna, Riyanka
    Pakray, Partha
    Bandyopadhyay, Sivaji
    COMPUTACION Y SISTEMAS, 2022, 26 (04): : 1669 - 1687
  • [3] Improved neural machine translation for low-resource English-Assamese pair
    Laskar, Sahinur Rahman
    Khilji, Abdullah Faiz Ur Rahman
    Pakray, Partha
    Bandyopadhyay, Sivaji
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 42 (05) : 4727 - 4738
  • [4] Improving neural machine translation by integrating transliteration for low-resource English-Assamese language
    Nath, Basab
    Sarkar, Sunita
    Mukhopadhyay, Somnath
    Roy, Arindam
    NATURAL LANGUAGE PROCESSING, 2025, 31 (02): : 306 - 327
  • [5] A Technique to Pre-trained Neural Network Language Model Customization to Software Development Domain
    Dudarin, Pavel, V
    Tronin, Vadim G.
    Svyatov, Kirill, V
    ARTIFICIAL INTELLIGENCE: (RCAI 2019), 2019, 1093 : 169 - 176
  • [6] Integrating Pre-Trained Language Model With Physical Layer Communications
    Lee, Ju-Hyung
    Lee, Dong-Ho
    Lee, Joohan
    Pujara, Jay
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2024, 23 (11) : 17266 - 17278
  • [7] Detection of Chinese Deceptive Reviews Based on Pre-Trained Language Model
    Weng, Chia-Hsien
    Lin, Kuan-Cheng
    Ying, Jia-Ching
    APPLIED SCIENCES-BASEL, 2022, 12 (07):
  • [8] PoliBERTweet: A Pre-trained Language Model for Analyzing Political Content on Twitter
    Kawintiranon, Kornraphop
    Singh, Lisa
    LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 7360 - 7367
  • [9] Addressing data scarcity issue for English-Mizo neural machine translation using data augmentation and language model
    Khenglawt, Vanlalmuansangi
    Laskar, Sahinur Rahman
    Pakray, Partha
    Khan, Ajoy Kumar
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2024, 46 (03) : 6313 - 6323
  • [10] Enriching Pre-trained Language Model with Entity Information for Relation Classification
    Wu, Shanchan
    He, Yifan
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 2361 - 2364