LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models

被引:0
|
作者
Hu, Zhiqiang [1 ]
Wang, Lei [2 ]
Lan, Yihuai
Xu, Wanyu [4 ]
Lim, Ee-Peng [2 ]
Bing, Lidong [3 ]
Xu, Xing [5 ]
Poria, Soujanya [1 ]
Lee, Roy Ka-Wei [1 ]
机构
[1] Singapore Univ Technol & Design, Singapore, Singapore
[2] Singapore Management Univ, Singapore, Singapore
[3] Alibaba Grp, DAMO Acad, Singapore, Singapore
[4] Southwest Jiaotong Univ, Chengdu, Peoples R China
[5] Univ Elect Sci & Technol China, Chengdu, Peoples R China
来源
2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The success of large language models (LLMs), like GPT-4 and ChatGPT, has led to the development of numerous cost-effective and accessible alternatives that are created by finetuning open-access LLMs with task-specific data (e.g., ChatDoctor) or instruction data (e.g., Alpaca). Among the various fine-tuning methods, adapter-based parameter-efficient fine-tuning (PEFT) is undoubtedly one of the most attractive topics, as it only requires fine-tuning a few external parameters instead of the entire LLMs while achieving comparable or even better performance. To enable further research on PEFT methods of LLMs, this paper presents LLM-Adapters, an easy-to-use framework that integrates various adapters into LLMs and can execute these adapter-based PEFT methods of LLMs for different tasks. The framework includes state-of-the-art open-access LLMs such as LLaMA, BLOOM, and GPT-J, as well as widely used adapters such as Series adapters, Parallel adapter, Prompt-based learning and Reparametrization-based methods. Moreover, we conduct extensive empirical studies on the impact of adapter types, placement locations, and hyper-parameters to the best design for each adapter-based methods. We evaluate the effectiveness of the adapters on fourteen datasets from two different reasoning tasks, Arithmetic Reasoning and Commonsense Reasoning. The results demonstrate that using adapter-based PEFT in smaller-scale LLMs (7B) with few extra trainable parameters yields comparable, and in some cases superior, performance to powerful LLMs (175B) in zero-shot inference on both reasoning tasks. The code and datasets can be found in https://github.com/AGI-Edgerunners/LLM-Adapters.
引用
收藏
页码:5254 / 5276
页数:23
相关论文
共 50 条
  • [31] One is Not Enough: Parameter-Efficient Fine-Tuning With Multiplicative Sparse Factorization
    Chen, Xuxi
    Chen, Tianlong
    Cheng, Yu
    Chen, Weizhu
    Awadallah, Ahmed Hassan
    Wang, Zhangyang
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2024, 18 (06) : 1059 - 1069
  • [32] Strong Baselines for Parameter-Efficient Few-Shot Fine-Tuning
    Basu, Samyadeep
    Hu, Shell
    Massiceti, Daniela
    Feizi, Soheil
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 11024 - 11031
  • [33] An Empirical Study of Parameter-Efficient Fine-Tuning Methods for Pre-trained Code Models
    Liu, Jiaxing
    Sha, Chaofeng
    Peng, Xin
    2023 38TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE, 2023, : 397 - 408
  • [34] ClassWise-SAM-Adapter: Parameter-Efficient Fine-Tuning Adapts Segment Anything to SAR Domain for Semantic Segmentation
    Pu, Xinyang
    Jia, Hecheng
    Zheng, Linghao
    Wang, Feng
    Xu, Feng
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2025, 18 : 4791 - 4804
  • [35] SCT: A Simple Baseline for Parameter-Efficient Fine-Tuning via Salient Channels
    Zhao, Henry Hengyuan
    Wang, Pichao
    Zhao, Yuyang
    Luo, Hao
    Wang, Fan
    Shou, Mike Zheng
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (03) : 731 - 749
  • [36] Federated Learning of Large Language Models with Parameter-Efficient Prompt Tuning and Adaptive Optimization
    Che, Tianshi
    Liu, Ji
    Zhou, Yang
    Ren, Jiaxiang
    Zhou, Jiwen
    Sheng, Victor S.
    Dai, Huaiyu
    Dou, Dejing
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 7871 - 7888
  • [37] Parameter-Efficient Fine-Tuning Method for Task-Oriented Dialogue Systems
    Mo, Yunho
    Yoo, Joon
    Kang, Sangwoo
    MATHEMATICS, 2023, 11 (14)
  • [38] Towards Robust and Generalized Parameter-Efficient Fine-Tuning for Noisy Label Learning
    Kim, Yeachan
    Kim, Junho
    Lee, SangKeun
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 5922 - 5936
  • [39] Towards Foundation Models and Few-Shot Parameter-Efficient Fine-Tuning for Volumetric Organ Segmentation
    Silva-Rodriguez, Julio
    Dolz, Jose
    Ben Ayed, Ismail
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023 WORKSHOPS, 2023, 14393 : 213 - 224
  • [40] Shadclips: When Parameter-Efficient Fine-Tuning with Multimodal Meets Shadow Removal
    Zhang, Xiaofeng
    Gu, Chaochen
    Xu, Zishan
    Tang, Hao
    Cheng, Hao
    Wu, Kaijie
    Zhu, Shanying
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2024, 38 (16)