Toward Low-Resource Languages Machine Translation: A Language-Specific Fine-Tuning With LoRA for Specialized Large Language Models

被引:0
作者
Liang, Xiao [1 ,2 ]
Khaw, Yen-Min Jasmina [1 ]
Liew, Soung-Yue [3 ]
Tan, Tien-Ping [4 ]
Qin, Donghong [2 ]
机构
[1] Univ Tunku Abdul Rahman, Fac Informat & Commun Technol, Dept Comp Sci, Kampar 31900, Malaysia
[2] Guangxi Minzu Univ, Sch Artificial Intelligence, Nanning 530008, Peoples R China
[3] Univ Tunku Abdul Rahman, Fac Informat & Commun Technol, Dept Comp & Commun Technol, Kampar 31900, Malaysia
[4] Univ Sains Malaysia, Sch Comp Sci, George Town 11700, Malaysia
来源
IEEE ACCESS | 2025年 / 13卷
关键词
Machine translation; low-resource languages; large language models; parameter-efficient fine-tuning; LoRA;
D O I
10.1109/ACCESS.2025.3549795
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the field of computational linguistics, addressing machine translation (MT) challenges for low-resource languages remains crucial, as these languages often lack extensive data compared to high-resource languages. General large language models (LLMs), such as GPT-4 and Llama, primarily trained on monolingual corpora, face significant challenges in translating low-resource languages, often resulting in subpar translation quality. This study introduces Language-Specific Fine-Tuning with Low-rank adaptation (LSFTL), a method that enhances translation for low-resource languages by optimizing the multi-head attention and feed-forward networks of Transformer layers through low-rank matrix adaptation. LSFTL preserves the majority of the model parameters while selectively fine-tuning key components, thereby maintaining stability and enhancing translation quality. Experiments on non-English centered low-resource Asian languages demonstrated that LSFTL improved COMET scores by 1-3 points compared to specialized multilingual machine translation models. Additionally, LSFTL's parameter-efficient approach allows smaller models to achieve performance comparable to their larger counterparts, highlighting its significance in making machine translation systems more accessible and effective for low-resource languages.
引用
收藏
页码:46616 / 46626
页数:11
相关论文
共 50 条
  • [41] Leveraging Large Language Models Knowledge Enhancement Dual-Stage Fine-Tuning Framework for Recommendation
    Zeng, Biqing
    Shi, Hao
    Li, Yangyu
    Li, Ruizhe
    Deng, Huimin
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT II, NLPCC 2024, 2025, 15360 : 333 - 345
  • [42] Enhancing Chinese Essay Discourse Logic Evaluation Through Optimized Fine-Tuning of Large Language Models
    Song, Jinwang
    Song, Yanxin
    Zhou, Guangyu
    Fu, Wenhui
    Zhang, Kunli
    Zan, Hongying
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT V, NLPCC 2024, 2025, 15363 : 342 - 352
  • [43] Investigating the Predominance of Large Language Models in Low-Resource Bangla Language over Transformer Models for Hate Speech Detection: A Comparative Analysis
    Faria, Fatema Tuj Johora
    Baniata, Laith H.
    Kang, Sangwoo
    MATHEMATICS, 2024, 12 (23)
  • [44] Enhancing Machine-Generated Text Detection: Adversarial Fine-Tuning of Pre-Trained Language Models
    Hee Lee, Dong
    Jang, Beakcheol
    IEEE ACCESS, 2024, 12 : 65333 - 65340
  • [45] Low-resource knowledge graph completion based on knowledge distillation driven by large language models
    Hou, Wenlong
    Zhao, Weidong
    Jia, Ning
    Liu, Xianhui
    APPLIED SOFT COMPUTING, 2025, 169
  • [46] Research on fine-tuning strategies for text classification in the aquaculture domain by combining deep learning and large language models
    Zhenglin Li
    Sijia Zhang
    Peirong Cao
    Jiaqi Zhang
    Zongshi An
    Aquaculture International, 2025, 33 (4)
  • [47] From pre-training to fine-tuning: An in-depth analysis of Large Language Models in the biomedical domain
    Bonfigli, Agnese
    Bacco, Luca
    Merone, Mario
    Dell'Orletta, Felice
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2024, 157
  • [48] Sample Size Considerations for Fine-Tuning Large Language Models for Named Entity Recognition Tasks: Methodological Study
    Majdik, Zoltan P.
    Graham, S. Scott
    Edward, Jade C. Shiva
    Rodriguez, Sabrina N.
    Karnes, Martha S.
    Jensen, Jared T.
    Barbour, Joshua B.
    Rousseau, Justin F.
    JMIR AI, 2024, 3
  • [49] Eliciting analogical reasoning from language models in retrieval-augmented translation under low-resource scenarios
    Wang, Liyan
    Wloka, Bartholomaus
    Lepage, Yves
    NEUROCOMPUTING, 2025, 630
  • [50] Xai-driven knowledge distillation of large language models for efficient deployment on low-resource devices
    Cantini, Riccardo
    Orsino, Alessio
    Talia, Domenico
    JOURNAL OF BIG DATA, 2024, 11 (01)