LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models

被引:0
|
作者
Hu, Zhiqiang [1 ]
Wang, Lei [2 ]
Lan, Yihuai
Xu, Wanyu [4 ]
Lim, Ee-Peng [2 ]
Bing, Lidong [3 ]
Xu, Xing [5 ]
Poria, Soujanya [1 ]
Lee, Roy Ka-Wei [1 ]
机构
[1] Singapore Univ Technol & Design, Singapore, Singapore
[2] Singapore Management Univ, Singapore, Singapore
[3] Alibaba Grp, DAMO Acad, Singapore, Singapore
[4] Southwest Jiaotong Univ, Chengdu, Peoples R China
[5] Univ Elect Sci & Technol China, Chengdu, Peoples R China
来源
2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The success of large language models (LLMs), like GPT-4 and ChatGPT, has led to the development of numerous cost-effective and accessible alternatives that are created by finetuning open-access LLMs with task-specific data (e.g., ChatDoctor) or instruction data (e.g., Alpaca). Among the various fine-tuning methods, adapter-based parameter-efficient fine-tuning (PEFT) is undoubtedly one of the most attractive topics, as it only requires fine-tuning a few external parameters instead of the entire LLMs while achieving comparable or even better performance. To enable further research on PEFT methods of LLMs, this paper presents LLM-Adapters, an easy-to-use framework that integrates various adapters into LLMs and can execute these adapter-based PEFT methods of LLMs for different tasks. The framework includes state-of-the-art open-access LLMs such as LLaMA, BLOOM, and GPT-J, as well as widely used adapters such as Series adapters, Parallel adapter, Prompt-based learning and Reparametrization-based methods. Moreover, we conduct extensive empirical studies on the impact of adapter types, placement locations, and hyper-parameters to the best design for each adapter-based methods. We evaluate the effectiveness of the adapters on fourteen datasets from two different reasoning tasks, Arithmetic Reasoning and Commonsense Reasoning. The results demonstrate that using adapter-based PEFT in smaller-scale LLMs (7B) with few extra trainable parameters yields comparable, and in some cases superior, performance to powerful LLMs (175B) in zero-shot inference on both reasoning tasks. The code and datasets can be found in https://github.com/AGI-Edgerunners/LLM-Adapters.
引用
收藏
页码:5254 / 5276
页数:23
相关论文
共 50 条
  • [41] SCT: A Simple Baseline for Parameter-Efficient Fine-Tuning via Salient Channels
    Henry Hengyuan Zhao
    Pichao Wang
    Yuyang Zhao
    Hao Luo
    Fan Wang
    Mike Zheng Shou
    International Journal of Computer Vision, 2024, 132 : 731 - 749
  • [42] Introducing Routing Functions to Vision-Language Parameter-Efficient Fine-Tuning with Low-Rank Bottlenecks
    Qui, Tingyu
    Tuytelaars, Tinne
    Moens, Marie-Francine
    COMPUTER VISION - ECCV 2024, PT LXXXVIII, 2025, 15146 : 291 - 308
  • [43] Phased Instruction Fine-Tuning for Large Language Models
    Pang, Wei
    Zhou, Chuan
    Zhou, Xiao-Hua
    Wang, Xiaojie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 5735 - 5748
  • [44] HackMentor: Fine-Tuning Large Language Models for Cybersecurity
    Zhang, Jie
    Wen, Hui
    Deng, Liting
    Xin, Mingfeng
    Li, Zhi
    Li, Lun
    Zhu, Hongsong
    Sun, Limin
    2023 IEEE 22ND INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, BIGDATASE, CSE, EUC, ISCI 2023, 2024, : 452 - 461
  • [45] Point-PEFT: Parameter-Efficient Fine-Tuning for 3D Pre-trained Models
    Tang, Yiwen
    Zhang, Ray
    Guo, Zoey
    Ma, Xianzheng
    Zhao, Bin
    Wang, Zhigang
    Wang, Dong
    Li, Xuelong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 5171 - 5179
  • [46] Parameter-efficient Multi-task Fine-tuning for Transformers via Shared Hypernetworks
    Mahabadi, Rabeeh Karimi
    Ruder, Sebastian
    Dehghani, Mostafa
    Henderson, James
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 565 - 576
  • [47] Data race detection via few-shot parameter-efficient fine-tuning
    Shen, Yuanyuan
    Peng, Manman
    Zhang, Fan
    Wu, Qiang
    JOURNAL OF SYSTEMS AND SOFTWARE, 2025, 222
  • [48] Structure-Aware Low-Rank Adaptation for Parameter-Efficient Fine-Tuning
    Hu, Yahao
    Xie, Yifei
    Wang, Tianfeng
    Chen, Man
    Pan, Zhisong
    MATHEMATICS, 2023, 11 (20)
  • [49] Parameter-efficient fine-tuning of pre-trained code models for just-in-time defect prediction
    Abu Talib M.
    Bou Nassif A.
    Azzeh M.
    Alesh Y.
    Afadar Y.
    Neural Computing and Applications, 36 (27) : 16911 - 16940
  • [50] UPetu: A Unified Parameter-Efficient Fine-Tuning Framework for Remote Sensing Foundation Model
    Dong, Zhe
    Gu, Yanfeng
    Liu, Tianzhu
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62 : 1 - 13