Toward Low-Resource Languages Machine Translation: A Language-Specific Fine-Tuning With LoRA for Specialized Large Language Models

被引:0
|
作者
Liang, Xiao [1 ,2 ]
Khaw, Yen-Min Jasmina [1 ]
Liew, Soung-Yue [3 ]
Tan, Tien-Ping [4 ]
Qin, Donghong [2 ]
机构
[1] Univ Tunku Abdul Rahman, Fac Informat & Commun Technol, Dept Comp Sci, Kampar 31900, Malaysia
[2] Guangxi Minzu Univ, Sch Artificial Intelligence, Nanning 530008, Peoples R China
[3] Univ Tunku Abdul Rahman, Fac Informat & Commun Technol, Dept Comp & Commun Technol, Kampar 31900, Malaysia
[4] Univ Sains Malaysia, Sch Comp Sci, George Town 11700, Malaysia
来源
IEEE ACCESS | 2025年 / 13卷
关键词
Machine translation; low-resource languages; large language models; parameter-efficient fine-tuning; LoRA;
D O I
10.1109/ACCESS.2025.3549795
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the field of computational linguistics, addressing machine translation (MT) challenges for low-resource languages remains crucial, as these languages often lack extensive data compared to high-resource languages. General large language models (LLMs), such as GPT-4 and Llama, primarily trained on monolingual corpora, face significant challenges in translating low-resource languages, often resulting in subpar translation quality. This study introduces Language-Specific Fine-Tuning with Low-rank adaptation (LSFTL), a method that enhances translation for low-resource languages by optimizing the multi-head attention and feed-forward networks of Transformer layers through low-rank matrix adaptation. LSFTL preserves the majority of the model parameters while selectively fine-tuning key components, thereby maintaining stability and enhancing translation quality. Experiments on non-English centered low-resource Asian languages demonstrated that LSFTL improved COMET scores by 1-3 points compared to specialized multilingual machine translation models. Additionally, LSFTL's parameter-efficient approach allows smaller models to achieve performance comparable to their larger counterparts, highlighting its significance in making machine translation systems more accessible and effective for low-resource languages.
引用
收藏
页码:46616 / 46626
页数:11
相关论文
共 50 条
  • [21] Selective privacy-preserving framework for large language models fine-tuning
    Wang, Teng
    Zhai, Lindong
    Yang, Tengfei
    Luo, Zhucheng
    Liu, Shuanggen
    INFORMATION SCIENCES, 2024, 678
  • [22] Extremely Low-resource Multilingual Neural Machine Translation for Indic Mizo Language
    Lalrempuii C.
    Soni B.
    International Journal of Information Technology, 2023, 15 (8) : 4275 - 4282
  • [23] Efficient fine-tuning of large language models for automated building energy modeling in complex cases
    Jiang, Gang
    Chen, Jianli
    AUTOMATION IN CONSTRUCTION, 2025, 175
  • [24] Resource-Efficient Fine-Tuning Strategies for Automatic MOS Prediction in Text-to-Speech for Low-Resource Languages
    Do, Phat
    Coler, Matt
    Dijkstra, Jelske
    Klabbers, Esther
    INTERSPEECH 2023, 2023, : 5466 - 5470
  • [25] Morpheme-Based Neural Machine Translation Models for Low-Resource Fusion Languages
    Gezmu, Andargachew Mekonnen
    Nuenberger, Andreas
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (09)
  • [26] Enhancing Machine Translation Across Multiple Domains and Languages with Large Language Models
    Lu, Hao
    Zhang, Rui
    Huang, Hui
    Song, Fuhai
    Liu, Junkai
    Ye, Yican
    Lang, Lang
    Zhao, Ziqing
    Yang, Muyun
    Cong, Rui
    MACHINE TRANSLATION, CCMT 2024, 2025, 2365 : 69 - 81
  • [27] Prompt Engineering or Fine-Tuning? A Case Study on Phishing Detection with Large Language Models
    Trad, Fouad
    Chehab, Ali
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2024, 6 (01): : 367 - 384
  • [28] Efficient Fine-Tuning Large Language Models for Knowledge-Aware Response Planning
    Minh Nguyen
    Kishan, K. C.
    Toan Nguyen
    Chadha, Ankit
    Thuy Vu
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT II, 2023, 14170 : 593 - 611
  • [29] Exploring ASR Models in Low-Resource Languages: Use-Case the Macedonian Language
    Bogdanoski, Konstantin
    Mishev, Kostadin
    Simjanoska, Monika
    Trajanov, Dimitar
    DEEP LEARNING THEORY AND APPLICATIONS, DELTA 2023, 2023, 1875 : 254 - 268
  • [30] Fine-Tuning Large Language Models to Improve Accuracy and Comprehensibility of Automated Code Review
    Yu, Yongda
    Rong, Guoping
    Shen, Haifeng
    Zhang, He
    Shao, Dong
    Wang, Min
    Wei, Zhao
    Xu, Yong
    Wang, Juhong
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2025, 34 (01)