LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models

被引:0
|
作者
Hu, Zhiqiang [1 ]
Wang, Lei [2 ]
Lan, Yihuai
Xu, Wanyu [4 ]
Lim, Ee-Peng [2 ]
Bing, Lidong [3 ]
Xu, Xing [5 ]
Poria, Soujanya [1 ]
Lee, Roy Ka-Wei [1 ]
机构
[1] Singapore Univ Technol & Design, Singapore, Singapore
[2] Singapore Management Univ, Singapore, Singapore
[3] Alibaba Grp, DAMO Acad, Singapore, Singapore
[4] Southwest Jiaotong Univ, Chengdu, Peoples R China
[5] Univ Elect Sci & Technol China, Chengdu, Peoples R China
来源
2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The success of large language models (LLMs), like GPT-4 and ChatGPT, has led to the development of numerous cost-effective and accessible alternatives that are created by finetuning open-access LLMs with task-specific data (e.g., ChatDoctor) or instruction data (e.g., Alpaca). Among the various fine-tuning methods, adapter-based parameter-efficient fine-tuning (PEFT) is undoubtedly one of the most attractive topics, as it only requires fine-tuning a few external parameters instead of the entire LLMs while achieving comparable or even better performance. To enable further research on PEFT methods of LLMs, this paper presents LLM-Adapters, an easy-to-use framework that integrates various adapters into LLMs and can execute these adapter-based PEFT methods of LLMs for different tasks. The framework includes state-of-the-art open-access LLMs such as LLaMA, BLOOM, and GPT-J, as well as widely used adapters such as Series adapters, Parallel adapter, Prompt-based learning and Reparametrization-based methods. Moreover, we conduct extensive empirical studies on the impact of adapter types, placement locations, and hyper-parameters to the best design for each adapter-based methods. We evaluate the effectiveness of the adapters on fourteen datasets from two different reasoning tasks, Arithmetic Reasoning and Commonsense Reasoning. The results demonstrate that using adapter-based PEFT in smaller-scale LLMs (7B) with few extra trainable parameters yields comparable, and in some cases superior, performance to powerful LLMs (175B) in zero-shot inference on both reasoning tasks. The code and datasets can be found in https://github.com/AGI-Edgerunners/LLM-Adapters.
引用
收藏
页码:5254 / 5276
页数:23
相关论文
共 50 条
  • [21] Transferrable DP-Adapter Tuning: A Privacy-Preserving Multimodal Parameter-Efficient Fine-Tuning Framework
    Ji, Lixia
    Xiao, Shijie
    Xu, Bingzhi
    Zhang, Han
    2024 IEEE 24TH INTERNATIONAL CONFERENCE ON SOFTWARE QUALITY, RELIABILITY AND SECURITY, QRS, 2024, : 471 - 482
  • [22] Hadamard Adapter: An Extreme Parameter-Efficient Adapter Tuning Method for Pre-trained Language Models
    Chen, Yuyan
    Fu, Qiang
    Fan, Ge
    Du, Lun
    Lou, Jian-Guang
    Han, Shi
    Zhang, Dongmei
    Li, Zhixu
    Xiao, Yanghua
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 276 - 285
  • [23] Parameter-Efficient Fine-Tuning without Introducing New Latency
    Liao, Baohao
    Meng, Yan
    Monz, Christof
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 4242 - 4260
  • [24] AutoPEFT : Automatic Configuration Search for Parameter-Efficient Fine-Tuning
    Zhou, Han
    Wan, Xingchen
    Vulic, Ivan
    Korhonen, Anna
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 525 - 542
  • [25] Sensitivity-Aware Visual Parameter-Efficient Fine-Tuning
    He, Haoyu
    Cai, Jianfei
    Zhang, Jing
    Tao, Dacheng
    Zhuang, Bohan
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11791 - 11801
  • [26] AdapterGNN: Parameter-Efficient Fine-Tuning Improves Generalization in GNNs
    Li, Shengrui
    Han, Xueting
    Bai, Jing
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 12, 2024, : 13600 - 13608
  • [27] Leveraging Parameter-Efficient Fine-Tuning for Multilingual Abstractive Summarization
    Shen, Jialun
    Wang, Yusong
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT III, NLPCC 2024, 2025, 15361 : 293 - 303
  • [28] Parameter-efficient fine-tuning for single image snow removal
    Dai, Xinwei
    Zhou, Yuanbo
    Qiu, Xintao
    Tang, Hui
    Tong, Tong
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 265
  • [29] Bayesian Parameter-Efficient Fine-Tuning for Overcoming Catastrophic Forgetting
    Chen, Haolin
    Garner, Philip N.
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 4253 - 4262
  • [30] A survey of efficient fine-tuning methods for Vision-Language Models - Prompt and Adapter
    Xing, Jialu
    Liu, Jianping
    Wang, Jian
    Sun, Lulu
    Chen, Xi
    Gu, Xunxun
    Wang, Yingfei
    COMPUTERS & GRAPHICS-UK, 2024, 119