Efficient Fine-Tuning Large Language Models for Knowledge-Aware Response Planning

被引:3
|
作者
Minh Nguyen [1 ]
Kishan, K. C. [2 ]
Toan Nguyen [2 ]
Chadha, Ankit [2 ]
Thuy Vu [2 ]
机构
[1] Univ Oregon, Dept Comp Sci, Eugene, OR USA
[2] Amazon Alexa AI, Palo Alto, CA 94301 USA
来源
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT II | 2023年 / 14170卷
关键词
Knowledge-Aware Response Planning; Question Answering; Large Language Models; Fine-tuning;
D O I
10.1007/978-3-031-43415-0_35
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large Language Models (LLMs) have shown impressive emergent language capabilities, especially in applications with high ambiguity, such as language reasoning and knowledge consolidation. However, previous work explores the use of LLMs for acquiring information using either parametric or external knowledge, which might lead to serious issues such as hallucination. Toward solving these issues, we present a novel approach of knowledge-aware response planning (KARP) and propose a novel framework that employs (i) a knowledge retriever to obtain relevant information from web documents or databases for a given user query, and (ii) a robust fine-tuning strategy for LLMs to exploit the retrieved external knowledge for planning a final response. Experimental results show that our proposed framework can provide natural, concise answers for open-domain questions with high accuracy.
引用
收藏
页码:593 / 611
页数:19
相关论文
共 50 条
  • [1] Knowledge-Aware Code Generation with Large Language Models
    Huang, Tao
    Sun, Zhihong
    Jin, Zhi
    Li, Ge
    Lyu, Chen
    PROCEEDINGS 2024 32ND IEEE/ACM INTERNATIONAL CONFERENCE ON PROGRAM COMPREHENSION, ICPC 2024, 2024, : 52 - 63
  • [2] Parameter-efficient fine-tuning in large language models: a survey of methodologies
    Wang, Luping
    Chen, Sheng
    Jiang, Linnan
    Pan, Shu
    Cai, Runze
    Yang, Sen
    Yang, Fei
    ARTIFICIAL INTELLIGENCE REVIEW, 2025, 58 (08)
  • [3] Personalized Large Language Models through Parameter Efficient Fine-Tuning Techniques
    Braga, Marco
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 3076 - 3076
  • [4] HackMentor: Fine-Tuning Large Language Models for Cybersecurity
    Zhang, Jie
    Wen, Hui
    Deng, Liting
    Xin, Mingfeng
    Li, Zhi
    Li, Lun
    Zhu, Hongsong
    Sun, Limin
    2023 IEEE 22ND INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, BIGDATASE, CSE, EUC, ISCI 2023, 2024, : 452 - 461
  • [5] Repeatability of Fine-Tuning Large Language Models Illustrated Using QLoRA
    Alahmari, Saeed S.
    Hall, Lawrence O.
    Mouton, Peter R.
    Goldgof, Dmitry B.
    IEEE ACCESS, 2024, 12 : 153221 - 153231
  • [6] Characterizing Communication in Distributed Parameter-Efficient Fine-Tuning for Large Language Models
    Alnaasan, Nawras
    Huang, Horng-Ruey
    Shafi, Aamir
    Subramoni, Hari
    Panda, Dhabaleswar K.
    2024 IEEE SYMPOSIUM ON HIGH-PERFORMANCE INTERCONNECTS, HOTI 2024, 2024, : 11 - 19
  • [7] Getting it right: the limits of fine-tuning large language models
    Browning, Jacob
    ETHICS AND INFORMATION TECHNOLOGY, 2024, 26 (02)
  • [8] Evaluating the Adaptability of Large Language Models for Knowledge-aware Question and Answering
    Thakkar, Jay
    Kolekar, Suresh
    Gite, Shilpa
    Pradhan, Biswajeet
    Alamri, Abdullah
    INTERNATIONAL JOURNAL ON SMART SENSING AND INTELLIGENT SYSTEMS, 2024, 17 (01):
  • [9] Prompting or Fine-tuning? A Comparative Study of Large Language Models for Taxonomy Construction
    Chen, Boqi
    Yi, Fandi
    Varro, Daniel
    2023 ACM/IEEE INTERNATIONAL CONFERENCE ON MODEL DRIVEN ENGINEERING LANGUAGES AND SYSTEMS COMPANION, MODELS-C, 2023, : 588 - 596
  • [10] Selective privacy-preserving framework for large language models fine-tuning
    Wang, Teng
    Zhai, Lindong
    Yang, Tengfei
    Luo, Zhucheng
    Liu, Shuanggen
    INFORMATION SCIENCES, 2024, 678