Optimizing Chinese-to-English Translation Using Large Language Models

被引:0
作者
Huang, Donghao [1 ,2 ]
Wang, Zhaoxia [1 ]
机构
[1] Singapore Management Univ, Sch Comp & Informat Syst, Singapore, Singapore
[2] Mastercard, Res & Dev, Singapore, Singapore
来源
2025 IEEE SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE IN NATURAL LANGUAGE PROCESSING AND SOCIAL MEDIA, CI-NLPSOME | 2025年
关键词
Machine Translation; Large Language Models; Natural Language Processing; Chinese-English Translation; COMET Metric; Fine-tuning; Efficiency Analysis; Zero-shot Learning; Few-shot Learning;
D O I
10.1109/CI-NLPSOME64976.2025.10970768
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The advent of Large Language Models (LLMs) has significantly advanced Chinese-to-English translation tasks. However, the translation process remains challenging due to the substantial differences in syntax, semantics, and morphology between these two languages, despite notable achievements. This paper presents a comprehensive study on Chinese-to-English translation, evaluating the performance of various LLMs. We explore a range of open-source models, from 3.5 billion to 72 billion parameters, and OpenAI's latest models, across zero-shot, few-shot, and fine-tuned learning paradigms. Our analysis assesses translation quality using the COMET metric, reliability with the Translation Completeness Ratio (TCR), and efficiency via Characters per Second (CPS). The results highlight substantial trade-offs between model size, translation accuracy, and processing speed. Larger models tend to produce higher-quality translations, whereas smaller models offer greater efficiency. Fine-tuning significantly improves the performance of open-source LLMs, surpassing few-shot learning in both translation quality and processing speed. Proprietary models like GPT-4o exhibit consistent high performance without significant gains from fine-tuning. We emphasize the potential of fine-tuning with techniques like LoRA/QLoRA to optimize the balance between translation accuracy and computational efficiency, offering valuable insights for deploying LLMs in real-world translation scenarios.
引用
收藏
页数:7
相关论文
共 23 条
[1]  
Brown TB, 2020, ADV NEUR IN, V33
[2]   Seven Pillars for the Future of Artificial Intelligence [J].
Cambria, Erik ;
Mao, Rui ;
Chen, Melvin ;
Wang, Zhaoxia ;
Ho, Seng-Beng .
IEEE INTELLIGENT SYSTEMS, 2023, 38 (06) :62-69
[3]  
Conneau Alexis, 2020, P 58 ANN M ASS COMPU, DOI [DOI 10.48550/ARXIV.1911.02116, 10.18653/v1/2020.acl-main.747, DOI 10.18653/V1/2020.ACL-MAIN.747]
[4]  
Dettmers T, 2023, Arxiv, DOI [arXiv:2305.14314, 10.48550/arxiv.2305.14314, DOI 10.48550/ARXIV.2305.14314]
[5]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[6]  
Hu EJ, 2021, Arxiv, DOI [arXiv:2106.09685, DOI 10.48550/ARXIV.2106.09685]
[7]   Automating Maritime Risk Data Collection and Identification Leveraging Large Language Models [J].
Huang, Donghao ;
Fu, Xiuju ;
Yin, Xiaofeng ;
Pen, Haibo ;
Wang, Zhaoxia .
2024 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW, 2024, :433-439
[8]   Performance Analysis of Llama 2 Among Other LLMs [J].
Huang, Donghao ;
Hu, Zhenda ;
Wang, Zhaoxia .
2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, :1081-1085
[9]  
Kocmi T., 2022, P 7 C MACH TRANSL WM, P46
[10]  
Koehn P., 2009, on Statistical Machine Translation