Toward Low-Resource Languages Machine Translation: A Language-Specific Fine-Tuning With LoRA for Specialized Large Language Models

被引:0
作者
Liang, Xiao [1 ,2 ]
Khaw, Yen-Min Jasmina [1 ]
Liew, Soung-Yue [3 ]
Tan, Tien-Ping [4 ]
Qin, Donghong [2 ]
机构
[1] Univ Tunku Abdul Rahman, Fac Informat & Commun Technol, Dept Comp Sci, Kampar 31900, Malaysia
[2] Guangxi Minzu Univ, Sch Artificial Intelligence, Nanning 530008, Peoples R China
[3] Univ Tunku Abdul Rahman, Fac Informat & Commun Technol, Dept Comp & Commun Technol, Kampar 31900, Malaysia
[4] Univ Sains Malaysia, Sch Comp Sci, George Town 11700, Malaysia
来源
IEEE ACCESS | 2025年 / 13卷
关键词
Machine translation; low-resource languages; large language models; parameter-efficient fine-tuning; LoRA;
D O I
10.1109/ACCESS.2025.3549795
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the field of computational linguistics, addressing machine translation (MT) challenges for low-resource languages remains crucial, as these languages often lack extensive data compared to high-resource languages. General large language models (LLMs), such as GPT-4 and Llama, primarily trained on monolingual corpora, face significant challenges in translating low-resource languages, often resulting in subpar translation quality. This study introduces Language-Specific Fine-Tuning with Low-rank adaptation (LSFTL), a method that enhances translation for low-resource languages by optimizing the multi-head attention and feed-forward networks of Transformer layers through low-rank matrix adaptation. LSFTL preserves the majority of the model parameters while selectively fine-tuning key components, thereby maintaining stability and enhancing translation quality. Experiments on non-English centered low-resource Asian languages demonstrated that LSFTL improved COMET scores by 1-3 points compared to specialized multilingual machine translation models. Additionally, LSFTL's parameter-efficient approach allows smaller models to achieve performance comparable to their larger counterparts, highlighting its significance in making machine translation systems more accessible and effective for low-resource languages.
引用
收藏
页码:46616 / 46626
页数:11
相关论文
共 50 条
  • [31] Research on Fine-Tuning Optimization Strategies for Large Language Models in Tabular Data Processing
    Zhao, Xiaoyong
    Leng, Xingxin
    Wang, Lei
    Wang, Ningning
    BIOMIMETICS, 2024, 9 (11)
  • [32] Cross-Lingual Transfer of Large Language Model by Visually-Derived Supervision Toward Low-Resource Languages
    Muraoka, Masayasu
    Bhattacharjee, Bishwaranjan
    Merler, Michele
    Blackwood, Graeme
    Li, Yulong
    Zhao, Yang
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 3637 - 3646
  • [33] Generative-Adversarial Networks for Low-Resource Language Data Augmentation in Machine Translation
    Zeng, Linda
    2024 6TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING, ICNLP 2024, 2024, : 11 - 18
  • [34] Parameter-Efficient Fine-Tuning of Pre-trained Large Language Models for Financial Text Analysis
    Langa, Kelly
    Wang, Hairong
    Okuboyejo, Olaperi
    ARTIFICIAL INTELLIGENCE RESEARCH, SACAIR 2024, 2025, 2326 : 3 - 20
  • [35] EW-Tune: A Framework for Privately Fine-Tuning Large Language Models with Differential Privacy
    Behnia, Rouzbeh
    Ebrahimi, Mohammadreza
    Pacheco, Jason
    Padmanabhan, Balaji
    2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW, 2022, : 560 - 566
  • [36] Fine-tuning and prompt engineering for large language models-based code review automation
    Pornprasit, Chanathip
    Tantithamthavorn, Chakkrit
    INFORMATION AND SOFTWARE TECHNOLOGY, 2024, 175
  • [37] Reparameterization-Based Parameter-Efficient Fine-Tuning Methods for Large Language Models: A Systematic Survey
    Chen, Zezhou
    Liu, Zhaoxiang
    Wang, Kai
    Lian, Shiguo
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT III, NLPCC 2024, 2025, 15361 : 107 - 118
  • [38] Fine-Tuning Self-Supervised Multilingual Sequence-To-Sequence Models for Extremely Low-Resource NMT
    Thillainathan, Sarubi
    Ranathunga, Surangika
    Jayasena, Sanath
    MORATUWA ENGINEERING RESEARCH CONFERENCE (MERCON 2021) / 7TH INTERNATIONAL MULTIDISCIPLINARY ENGINEERING RESEARCH CONFERENCE, 2021, : 432 - 437
  • [39] Contrastive Learning for Morphological Disambiguation Using Large Language Models in Low-Resource Settings
    Tolegen, Gulmira
    Toleu, Alymzhan
    Mussabayev, Rustam
    APPLIED SCIENCES-BASEL, 2024, 14 (21):
  • [40] Leveraging Large Language Models Knowledge Enhancement Dual-Stage Fine-Tuning Framework for Recommendation
    Zeng, Biqing
    Shi, Hao
    Li, Yangyu
    Li, Ruizhe
    Deng, Huimin
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT II, NLPCC 2024, 2025, 15360 : 333 - 345