Improving Multilingual Neural Machine Translation System for Indic Languages

被引:4
|
作者
Das, Sudhansu Bala [1 ]
Biradar, Atharv [2 ]
Mishra, Tapas Kumar [1 ]
Patra, Bidyut Kr. [3 ]
机构
[1] Natl Inst Technol NIT, Rourkela 769008, Odisha, India
[2] Pune Inst Comp Technol PICT, Pune, Maharashtra, India
[3] Indian Inst Technol IIT, Varanasi, Uttar Pradesh, India
关键词
Multilingual neuralmachine translation system (MNMT); Indic languages (ILs); low resource language; corpus; BLEU score;
D O I
10.1145/3587932
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Machine Translation System (MTS) serves as effective tool for communication by translating text or speech from one language to another language. Recently, neural machine translation (NMT) has become popular for its performance and cost-effectiveness. However, NMT systems are restricted in translating low-resource languages as a huge quantity of data is required to learn useful mappings across languages. The need for an efficient translation system becomes obvious in a large multilingual environment like India. Indian languages (ILs) are still treated as low-resource languages due to unavailability of corpora. In order to address such an asymmetric nature, the multilingual neural machine translation (MNMT) system evolves as an ideal approach in this direction. The MNMT converts many languages using a single model, which is extremely useful in terms of training process and lowering online maintenance costs. It is also helpful for improving low-resource translation. In this article, we propose an MNMT system to address the issues related to low-resource language translation. Our model comprises two MNMT systems, i.e., for English-Indic (one-to-many) and for Indic-English (many-to-one) with a shared encoder-decoder containing 15 language pairs (30 translation directions). Since most of IL pairs have a scanty amount of parallel corpora, not sufficient for training any machine translation model, we explore various augmentation strategies to improve overall translation quality through the proposed model. A state-of-the-art transformer architecture is used to realize the proposed model. In addition, the article addresses the use of language relationships (in terms of dialect, script, etc.), particularly about the role of high-resource languages of the same family in boosting the performance of low-resource languages. Moreover, the experimental results also show the advantage of back-translation and domain adaptation for ILs to enhance the translation quality of both source and target languages. Using all these key approaches, our proposed model emerges to be more efficient than the baseline model in terms of evaluation metrics, i.e., BLEU (BiLingual Evaluation Understudy) score for a set of ILs.
引用
收藏
页数:24
相关论文
共 45 条
  • [21] Multimodal Neural Machine Translation for English-Assamese Pair
    Laskar, Sahinur Rahman
    Paul, Bishwaraj
    Paudwal, Siddharth
    Gautam, Pranjit
    Biswas, Nirmita
    Pakray, Partha
    2021 INTERNATIONAL CONFERENCE ON COMPUTATIONAL PERFORMANCE EVALUATION (COMPE-2021), 2021, : 387 - 392
  • [22] Statistical Machine Translation Algorithm Based on Improved Neural Network
    Bing, Hu
    2017 INTERNATIONAL CONFERENCE ON ROBOTS & INTELLIGENT SYSTEM (ICRIS), 2017, : 294 - 297
  • [23] English–Mizo Machine Translation using neural and statistical approaches
    Amarnath Pathak
    Partha Pakray
    Jereemi Bentham
    Neural Computing and Applications, 2019, 31 : 7615 - 7631
  • [24] Improving phrase-based statistical machine translation with morphosyntactic transformation
    Thai Phuong Nguyen
    Shimazu, Akira
    MACHINE TRANSLATION, 2006, 20 (03) : 147 - 166
  • [25] Optimization of data analysis models for low-resource Eurasian languages using machine translation
    Chen, Hongyan
    Yee, Kim Kyung
    INTERNET TECHNOLOGY LETTERS, 2024,
  • [26] English-Mizo Machine Translation using neural and statistical approaches
    Pathak, Amarnath
    Pakray, Partha
    Bentham, Jereemi
    NEURAL COMPUTING & APPLICATIONS, 2019, 31 (11) : 7615 - 7631
  • [27] Development of a Recurrent Neural Network Model for English to Yoruba Machine Translation
    Esan, Adebimpe
    Oladosu, John
    Oyeleye, Christopher
    Adeyanju, Ibrahim
    Olaniyan, Olatayo
    Okomba, Nnamdi
    Omodunbi, Bolaji
    Adanigbo, Opeyemi
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2020, 11 (05) : 602 - 609
  • [28] Neural machine translation for limited resources English-Nyishi pair
    Kakum, Nabam
    Laskar, Sahinur Rahman
    Sambyo, Koj
    Pakray, Partha
    SADHANA-ACADEMY PROCEEDINGS IN ENGINEERING SCIENCES, 2023, 48 (04):
  • [29] Development of a recurrent neural network model for english to yoruba machine translation
    Esan A.
    Oladosu J.
    Oyeleye C.
    Adeyanju I.
    Olaniyan O.
    Okomba N.
    Omodunbi B.
    Adanigbo O.
    International Journal of Advanced Computer Science and Applications, 2020, 11 (05): : 602 - 609
  • [30] Neural machine translation for limited resources English-Nyishi pair
    Nabam Kakum
    Sahinur Rahman Laskar
    Koj Sambyo
    Partha Pakray
    Sādhanā, 48