Structure-Aware Low-Rank Adaptation for Parameter-Efficient Fine-Tuning

被引:3
|
作者
Hu, Yahao [1 ]
Xie, Yifei [1 ]
Wang, Tianfeng [1 ]
Chen, Man [1 ]
Pan, Zhisong [1 ]
机构
[1] Army Engn Univ PLA, Command & Control Engn Coll, Nanjing 210007, Peoples R China
基金
中国国家自然科学基金;
关键词
pre-trained language models; parameter-efficient fine-tuning; low-rank adaptation; intrinsic rank; training efficiency;
D O I
10.3390/math11204317
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
With the growing scale of pre-trained language models (PLMs), full parameter fine-tuning becomes prohibitively expensive and practically infeasible. Therefore, parameter-efficient adaptation techniques for PLMs have been proposed to learn through incremental updates of pre-trained weights, such as in low-rank adaptation (LoRA). However, LoRA relies on heuristics to select the modules and layers to which it is applied, and assigns them the same rank. As a consequence, any fine-tuning that ignores the structural information between modules and layers is suboptimal. In this work, we propose structure-aware low-rank adaptation (SaLoRA), which adaptively learns the intrinsic rank of each incremental matrix by removing rank-0 components during training. We conduct comprehensive experiments using pre-trained models of different scales in both task-oriented (GLUE) and task-agnostic (Yelp and GYAFC) settings. The experimental results show that SaLoRA effectively captures the structure-aware intrinsic rank. Moreover, our method consistently outperforms LoRA without significantly compromising training efficiency.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] LoRAPrune: Structured Pruning Meets Low-Rank Parameter-Efficient Fine-Tuning
    Zhang, Mingyang
    Chen, Hao
    Shen, Chunhua
    Yang, Zhen
    Ou, Linlin
    Yu, Xinyi
    Zhuang, Bohan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 3013 - 3026
  • [2] Hydra: Multi-head low-rank adaptation for parameter efficient fine-tuning
    Kim, Sanghyeon
    Yang, Hyunmo
    Kim, Yunghyun
    Hong, Youngjoon
    Park, Eunbyung
    NEURAL NETWORKS, 2024, 178
  • [3] Introducing Routing Functions to Vision-Language Parameter-Efficient Fine-Tuning with Low-Rank Bottlenecks
    Qui, Tingyu
    Tuytelaars, Tinne
    Moens, Marie-Francine
    COMPUTER VISION - ECCV 2024, PT LXXXVIII, 2025, 15146 : 291 - 308
  • [4] Leveraging Low-Rank Adaptation for Parameter-Efficient Fine-Tuning in Multi-Speaker Adaptive Text-to-Speech Synthesis
    Hong, Changi
    Lee, Jung Hyuk
    Kim, Hong Kook
    IEEE ACCESS, 2024, 12 : 190711 - 190727
  • [5] Sensitivity-Aware Visual Parameter-Efficient Fine-Tuning
    He, Haoyu
    Cai, Jianfei
    Zhang, Jing
    Tao, Dacheng
    Zhuang, Bohan
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11791 - 11801
  • [6] On the Effectiveness of Parameter-Efficient Fine-Tuning
    Fu, Zihao
    Yang, Haoran
    So, Anthony Man-Cho
    Lam, Wai
    Bing, Lidong
    Collier, Nigel
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 12799 - 12807
  • [7] Dropout Mixture Low-Rank Adaptation for Visual Parameters-Efficient Fine-Tuning
    Fang, Zhengyi
    Wang, Yue
    Yi, Ran
    Ma, Lizhuang
    COMPUTER VISION-ECCV 2024, PT VII, 2025, 15065 : 369 - 386
  • [8] Pass-Tuning: Towards Structure-Aware Parameter-Efficient Tuning for Code Representation Learning
    Chen, Nuo
    Sun, Qiushi
    Wang, Jianing
    Li, Xiang
    Gao, Ming
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 577 - 591
  • [9] Frozen Weights as Prior for Parameter-Efficient Fine-Tuning
    Ma, Xiaolong
    Liu, Peishun
    Gao, Haojie
    Yan, Zikang
    Ma, Ningning
    Liu, Wenqiang
    Wang, Xuefang
    Tang, Ruichun
    IEEE ACCESS, 2025, 13 : 24411 - 24425
  • [10] AFLoRA: Adaptive Freezing of Low Rank Adaptation in Parameter Efficient Fine-Tuning of Large Models
    Lin, Zeyu
    Kundu, Souvik
    Li, Anni
    Wan, Junrui
    Jiang, Lianghao
    Beerell, Peter A.
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 161 - 167