Performance comparison of retrieval-augmented generation and fine-tuned large language models for construction safety management knowledge retrieval

被引:3
|
作者
Lee, Jungwon [1 ]
Ahn, Seungjun [1 ]
Kim, Daeho [2 ]
Kim, Dongkyun [1 ]
机构
[1] Hongik Univ, Dept Civil & Environm Engn, Seoul, South Korea
[2] Univ Toronto, Dept Civil & Mineral Engn, Toronto, ON, Canada
基金
新加坡国家研究基金会;
关键词
Large Language Model (LLM); Retrieval-Augmented Generation (RAG); Fine-tuned LLM; Construction safety; Knowledge graph;
D O I
10.1016/j.autcon.2024.105846
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Construction safety standards are in unstructured formats like text and images, complicating their effective use in daily tasks. This paper compares the performance of Retrieval-Augmented Generation (RAG) and fine-tuned Large Language Model (LLM) for the construction safety knowledge retrieval. The RAG model was created by integrating GPT-4 with a knowledge graph derived from construction safety guidelines, while the fine-tuned LLM was fine-tuned using a question-answering dataset derived from the same guidelines. These models' performance is tested through case studies, using accident synopses as a query to generate preventive measurements. The responses were assessed using metrics, including cosine similarity, Euclidean distance, BLEU, and ROUGE scores. It was found that both models outperformed GPT-4, with the RAG model improving by 21.5 % and the fine-tuned LLM by 26 %. The findings highlight the relative strengths and weaknesses of the RAG and fine-tuned LLM approaches in terms of applicability and reliability for safety management.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Benchmarking Large Language Models in Retrieval-Augmented Generation
    Chen, Jiawei
    Lin, Hongyu
    Han, Xianpei
    Sun, Le
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 17754 - 17762
  • [2] BASHEXPLAINER: Retrieval-Augmented Bash Code Comment Generation based on Fine-tuned CodeBERT
    Yu, Chi
    Yang, Guang
    Chen, Xiang
    Liu, Ke
    Zhou, Yanlin
    2022 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE MAINTENANCE AND EVOLUTION (ICSME 2022), 2022, : 82 - 93
  • [3] Enhancing Retrieval-Augmented Large Language Models with Iterative Retrieval-Generation Synergy
    Shaol, Zhihong
    Gong, Yeyun
    Shen, Yelong
    Huang, Minlie
    Duane, Nan
    Chen, Weizhu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 9248 - 9274
  • [4] Application of retrieval-augmented generation for interactive industrial knowledge management via a large language model
    Chen, Lun-Chi
    Pardeshi, Mayuresh Sunil
    Liao, Yi-Xiang
    Pai, Kai-Chih
    COMPUTER STANDARDS & INTERFACES, 2025, 94
  • [5] Query Rewriting for Retrieval-Augmented Large Language Models
    Ma, Xinbei
    Gong, Yeyun
    He, Pengcheng
    Zhao, Hai
    Duan, Nan
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 5303 - 5315
  • [6] Integrating Graph Retrieval-Augmented Generation With Large Language Models for Supplier Discovery
    Li, Yunqing
    Ko, Hyunwoong
    Ameri, Farhad
    JOURNAL OF COMPUTING AND INFORMATION SCIENCE IN ENGINEERING, 2025, 25 (02)
  • [7] TrojanRAG: Retrieval-Augmented Generation Can Be Backdoor Driver in Large Language Models
    Shanghai Jiao Tong University, China
    arXiv,
  • [8] Hallucination Mitigation for Retrieval-Augmented Large Language Models: A Review
    Zhang, Wan
    Zhang, Jing
    MATHEMATICS, 2025, 13 (05)
  • [9] Enhancement of the Performance of Large Language Models inDiabetes Education through Retrieval-Augmented Generation:Comparative Study
    Wang, Dingqiao
    Liang, Jiangbo
    Ye, Jinguo
    Li, Jingni
    Li, Jingpeng
    Zhang, Qikai
    Hu, Qiuling
    Pan, Caineng
    Wang, Dongliang
    Liu, Zhong
    Shi, Wen
    Shi, Danli
    Li, Fei
    Qu, Bo
    Zheng, Yingfeng
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2024, 26
  • [10] Resolving Unseen Rumors with Retrieval-Augmented Large Language Models
    Chen, Lei
    Wei, Zhongyu
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT IV, NLPCC 2024, 2025, 15362 : 319 - 332