Structure-Aware Low-Rank Adaptation for Parameter-Efficient Fine-Tuning

被引:3
|
作者
Hu, Yahao [1 ]
Xie, Yifei [1 ]
Wang, Tianfeng [1 ]
Chen, Man [1 ]
Pan, Zhisong [1 ]
机构
[1] Army Engn Univ PLA, Command & Control Engn Coll, Nanjing 210007, Peoples R China
基金
中国国家自然科学基金;
关键词
pre-trained language models; parameter-efficient fine-tuning; low-rank adaptation; intrinsic rank; training efficiency;
D O I
10.3390/math11204317
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
With the growing scale of pre-trained language models (PLMs), full parameter fine-tuning becomes prohibitively expensive and practically infeasible. Therefore, parameter-efficient adaptation techniques for PLMs have been proposed to learn through incremental updates of pre-trained weights, such as in low-rank adaptation (LoRA). However, LoRA relies on heuristics to select the modules and layers to which it is applied, and assigns them the same rank. As a consequence, any fine-tuning that ignores the structural information between modules and layers is suboptimal. In this work, we propose structure-aware low-rank adaptation (SaLoRA), which adaptively learns the intrinsic rank of each incremental matrix by removing rank-0 components during training. We conduct comprehensive experiments using pre-trained models of different scales in both task-oriented (GLUE) and task-agnostic (Yelp and GYAFC) settings. The experimental results show that SaLoRA effectively captures the structure-aware intrinsic rank. Moreover, our method consistently outperforms LoRA without significantly compromising training efficiency.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] One is Not Enough: Parameter-Efficient Fine-Tuning With Multiplicative Sparse Factorization
    Chen, Xuxi
    Chen, Tianlong
    Cheng, Yu
    Chen, Weizhu
    Awadallah, Ahmed Hassan
    Wang, Zhangyang
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2024, 18 (06) : 1059 - 1069
  • [22] Parameter-efficient fine-tuning in large language models: a survey of methodologies
    Luping Wang
    Sheng Chen
    Linnan Jiang
    Shu Pan
    Runze Cai
    Sen Yang
    Fei Yang
    Artificial Intelligence Review, 58 (8)
  • [23] PELA: Learning Parameter-Efficient Models with Low-Rank Approximation
    Guo, Yangyang
    Wang, Guangzhi
    Kankanhalli, Mohan
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 15699 - 15709
  • [24] Strong Baselines for Parameter-Efficient Few-Shot Fine-Tuning
    Basu, Samyadeep
    Hu, Shell
    Massiceti, Daniela
    Feizi, Soheil
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 11024 - 11031
  • [25] Parameter-Efficient Fine-Tuning Large Speech Model Based on LoRA
    Ou, Ling
    Feng, Gen
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 36 - 41
  • [26] Parameter-efficient fine-tuning of large language models using semantic knowledge tuning
    Prottasha, Nusrat Jahan
    Mahmud, Asif
    Sobuj, Md. Shohanur Islam
    Bhat, Prakash
    Kowsher, Md
    Yousefi, Niloofar
    Garibay, Ozlem Ozmen
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [27] Efficient Fine-Tuning of Large Language Models via a Low-Rank Gradient Estimator
    Zhang, Luoming
    Lou, Zhenyu
    Ying, Yangwei
    Yang, Cheng
    Zhou, Hong
    APPLIED SCIENCES-BASEL, 2025, 15 (01):
  • [28] SCT: A Simple Baseline for Parameter-Efficient Fine-Tuning via Salient Channels
    Zhao, Henry Hengyuan
    Wang, Pichao
    Zhao, Yuyang
    Luo, Hao
    Wang, Fan
    Shou, Mike Zheng
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (03) : 731 - 749
  • [29] Parameter-Efficient Fine-Tuning Method for Task-Oriented Dialogue Systems
    Mo, Yunho
    Yoo, Joon
    Kang, Sangwoo
    MATHEMATICS, 2023, 11 (14)
  • [30] Towards Robust and Generalized Parameter-Efficient Fine-Tuning for Noisy Label Learning
    Kim, Yeachan
    Kim, Junho
    Lee, SangKeun
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 5922 - 5936