Large language model for interpreting research policy using adaptive two-stage retrieval augmented fine-tuning method

被引:0
|
作者
Ren, Runtao [1 ]
Ma, Jian [1 ]
Zheng, Zhimin [2 ]
机构
[1] City Univ Hong Kong, Dept Informat Syst, Kowloon Tong, Hong Kong, Peoples R China
[2] Natl Nat Sci Fdn China, Bur Planning, Beijing, Peoples R China
关键词
Generative AI; Large Language Model; Retrieval-augmented Generation; Fine-tuning; Interpretability;
D O I
10.1016/j.eswa.2025.127330
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Accurate interpretation of scientific funding policies is crucial for government funding agencies and research institutions to make informed decisions and allocate research funds effectively. However, current large language model (LLM)-based systems often generate responses without references, leading to a lack of interpretability needed for policy enforcement. This study introduces the Adaptive Two-stage Retrieval Augmented Fine-Tuning (AT-RAFT) method, a novel LLM-based approach specifically designed for science policy interpretation. AT-RAFT incorporates three complementary artifacts: a two-stage retrieval mechanism, adaptive hard-negative fine-tuning, and an interpretable response interface. It is trained directly on policy documents, allowing the model to provide reference answers based on retrieved text while also offering the original policy context to enhance interpretability. Our experiments demonstrate that AT-RAFT improves retrieval accuracy by 48% and generation performance by 44% compared to existing baseline systems, effectively supporting real-world decision-making tasks for stakeholders in research institutions and funding agencies. Our proposed method has been adopted by ScholarMate, the largest professional research social networking platform in China, and is now deployed on their platform, providing global users with access to advanced policy interpretation tools. Additionally, a demo version of the instantiated interface is available at https://github.com/renruntao/ResearchPolicy_RAG.
引用
收藏
页数:16
相关论文
共 26 条
  • [1] Unveiling the Power of Large Language Models: A Comparative Study of Retrieval-Augmented Generation, Fine-Tuning, and Their Synergistic Fusion for Enhanced Performance
    Budakoglu, Gulsum
    Emekci, Hakan
    IEEE ACCESS, 2025, 13 : 30936 - 30951
  • [2] A two-stage fine-tuning method for low-resource cross-lingual summarization
    Zhang, Kaixiong
    Zhang, Yongbing
    Yu, Zhengtao
    Huang, Yuxin
    Tan, Kaiwen
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2024, 21 (01) : 1125 - 1143
  • [3] Fine-Tuning Retrieval-Augmented Generation with an Auto-Regressive Language Model for Sentiment Analysis in Financial Reviews
    Mathebula, Miehleketo
    Modupe, Abiodun
    Marivate, Vukosi
    APPLIED SCIENCES-BASEL, 2024, 14 (23):
  • [4] Repeatability of Fine-Tuning Large Language Models Illustrated Using QLoRA
    Alahmari, Saeed S.
    Hall, Lawrence O.
    Mouton, Peter R.
    Goldgof, Dmitry B.
    IEEE ACCESS, 2024, 12 : 153221 - 153231
  • [5] Efficient fine-tuning of short text classification based on large language model
    Wang, Likun
    PROCEEDINGS OF INTERNATIONAL CONFERENCE ON MODELING, NATURAL LANGUAGE PROCESSING AND MACHINE LEARNING, CMNM 2024, 2024, : 33 - 38
  • [6] Fine-Tuning a Large Language Model with Reinforcement Learning for Educational Question Generation
    Lamsiyah, Salima
    El Mahdaouy, Abdelkader
    Nourbakhsh, Aria
    Schommer, Christoph
    ARTIFICIAL INTELLIGENCE IN EDUCATION, PT I, AIED 2024, 2024, 14829 : 424 - 438
  • [7] Research on Fine-Tuning Optimization Strategies for Large Language Models in Tabular Data Processing
    Zhao, Xiaoyong
    Leng, Xingxin
    Wang, Lei
    Wang, Ningning
    BIOMIMETICS, 2024, 9 (11)
  • [8] Evaluating the Effectiveness of Fine-Tuning Large Language Model for Domain-Specific Task
    Dabhi, Saumya
    Martinez, Joseph
    Poursardar, Faryaneh
    2024 IEEE INTERNATIONAL CONFERENCE ON INFORMATION REUSE AND INTEGRATION FOR DATA SCIENCE, IRI 2024, 2024, : 176 - 177
  • [9] VTT-LLM: Advancing Vulnerability-to-Tactic-and-Technique Mapping through Fine-Tuning of Large Language Model
    Zhang, Chenhui
    Wang, Le
    Fan, Dunqiu
    Zhu, Junyi
    Zhou, Tang
    Zeng, Liyi
    Li, Zhaohua
    MATHEMATICS, 2024, 12 (09)
  • [10] Receiver-Agnostic Radio Frequency Fingerprinting Based on Two-stage Unsupervised Domain Adaptation and Fine-tuning
    Bao, Jiazhong
    Xie, Xin
    Lu, Zhaoyi
    Hong, Jianan
    Hua, Cunqing
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 6085 - 6090