Unveiling the Power of Large Language Models: A Comparative Study of Retrieval-Augmented Generation, Fine-Tuning, and Their Synergistic Fusion for Enhanced Performance

被引:0
作者
Budakoglu, Gulsum [1 ]
Emekci, Hakan [1 ]
机构
[1] TED Univ, Grad Sch Appl Data Sci, TR-06420 Ankara, Turkiye
来源
IEEE ACCESS | 2025年 / 13卷
关键词
Retrieval augmented generation; Question answering (information retrieval); Adaptation models; Tuning; Hands; Accuracy; Training; Solid modeling; Online services; Large language models; Large language models (LLMs); retrieval-augmented generation (RAG); fine-tuning; hybrid models; performance optimization;
D O I
10.1109/ACCESS.2025.3542334
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Large-language model optimization for a particular application is crucial and challenging in natural language processing. This study compares two salient techniques for retrieve-augmented generation (RAG) and fine-tuning along with a new hybrid method that combines both. In this study, we investigate the effectiveness of various methods using the Stanford Question Answering Dataset (SQuAD), Microsoft Machine Reading Comprehension (MS MARCO) and SQL CREATE TABLE statements. RAG is used because it enriches the model responses with external data without much computational load during the inference. Fine-tuning updates the model parameters to improve the contextual accuracy. Our hybrid model balances the accuracy and efficiency of the two techniques. While fine-tuning entails semantic precision, RAG is more resource efficient. The hybrid approach while it may not offer surpassing results over fine-tuning-offers a balanced solution in scenarios where the application demands both efficiency and accuracy. These findings represent the trade-off involved in LLM optimization and offers a scope for further studies and practical applications.
引用
收藏
页码:30936 / 30951
页数:16
相关论文
共 39 条
  • [1] Alawwad HA, 2025, Arxiv, DOI arXiv:2402.05128
  • [2] Alghisi S, 2024, Arxiv, DOI arXiv:2406.06399
  • [3] Amil Merchant, 2020, arXiv
  • [4] [Anonymous], 2024, b-mc2/sql-create-context Datasets at Hugging Face
  • [5] Asai A, 2022, Arxiv, DOI arXiv:2211.09260
  • [6] Bajaj P, 2018, Arxiv, DOI arXiv:1611.09268
  • [7] Balaguer A, 2024, Arxiv, DOI [arXiv:2401.08406, DOI 10.48550/ARXIV.2401.08406, 10.48550/arXiv.2401.08406]
  • [8] Barnett S, 2024, Arxiv, DOI arXiv:2406.11201
  • [9] EW-Tune: A Framework for Privately Fine-Tuning Large Language Models with Differential Privacy
    Behnia, Rouzbeh
    Ebrahimi, Mohammadreza
    Pacheco, Jason
    Padmanabhan, Balaji
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW, 2022, : 560 - 566
  • [10] Brown TB, 2020, ADV NEUR IN, V33