PELA: Learning Parameter-Efficient Models with Low-Rank Approximation

被引:0
|
作者
Guo, Yangyang [1 ]
Wang, Guangzhi [1 ]
Kankanhalli, Mohan [1 ]
机构
[1] Natl Univ Singapore, Singapore, Singapore
来源
2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2024年
基金
新加坡国家研究基金会;
关键词
D O I
10.1109/CVPR52733.2024.01486
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Applying a pre-trained large model to downstream tasks is prohibitive under resource-constrained conditions. Re-cent dominant approaches for addressing efficiency issues involve adding a few learnable parameters to the fixed backbone model. This strategy, however, leads to more challenges in loading large models for downstream fine-tuning with limited resources. In this paper, we propose a novel method for increasing the parameter efficiency of pre-trained models by introducing an intermediate pre-training stage. To this end, we first employ low-rank approximation to compress the original large model and then devise a feature distillation module and a weight perturbation regularization module. These modules are specifically designed to enhance the low-rank model. In particular, we update only the low-rank model while freezing the backbone parameters during pre-training. This allows for direct and efficient utilization of the low-rank model for downstream fine-tuning tasks. The proposed method achieves both efficiencies in terms of required parameters and computation time while maintaining comparable results with minimal modifications to the backbone architecture. Specifically, when applied to three vision-only and one vision-language Transformer models, our approach often demonstrates a merely similar to 0.6 point decrease in performance while reducing the original parameter size by 1/3 to 2/3. We release our code at link.
引用
收藏
页码:15699 / 15709
页数:11
相关论文
共 50 条
  • [31] Joint Embedding Learning and Low-Rank Approximation: A Framework for Incomplete Multiview Learning
    Tao, Hong
    Hou, Chenping
    Yi, Dongyun
    Zhu, Jubo
    Hu, Dewen
    IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (03) : 1690 - 1703
  • [32] Multiscale Decomposition in Low-Rank Approximation
    Abdolali, Maryam
    Rahmati, Mohammad
    IEEE SIGNAL PROCESSING LETTERS, 2017, 24 (07) : 1015 - 1019
  • [33] SIMPLICIAL APPROXIMATION AND LOW-RANK TREES
    GILLET, H
    SHALEN, PB
    SKORA, RK
    COMMENTARII MATHEMATICI HELVETICI, 1991, 66 (04) : 521 - 540
  • [34] Enhanced Low-Rank Matrix Approximation
    Parekh, Ankit
    Selesnick, Ivan W.
    IEEE SIGNAL PROCESSING LETTERS, 2016, 23 (04) : 493 - 497
  • [35] Modifiable low-rank approximation to a matrix
    Barlow, Jesse L.
    Erbay, Hasan
    NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS, 2009, 16 (10) : 833 - 860
  • [36] Algorithms for lp Low-Rank Approximation
    Chierichetti, Flavio
    Gollapudi, Sreenivas
    Kumar, Ravi
    Lattanzi, Silvio
    Panigrahy, Rina
    Woodruff, David P.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [37] Low-Rank Matrix Approximation with Stability
    Li, Dongsheng
    Chen, Chao
    Lv, Qin
    Yan, Junchi
    Shang, Li
    Chu, Stephen M.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [38] Low-rank approximation of path-based traffic network models
    Zhang, Pengji
    Qian, Sean
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2025, 172
  • [39] Low-rank variance approximation in GMRF models: Single and multiscale approaches
    Malioutov, Dmitry M.
    Johnson, Jason K.
    Choi, Nlyung Jin
    Willsky, Alan S.
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2008, 56 (10) : 4621 - 4634
  • [40] Learning Nonlocal Sparse and Low-Rank Models for Image Compressive Sensing: Nonlocal sparse and low-rank modeling
    Zha, Zhiyuan
    Wen, Bihan
    Yuan, Xin
    Ravishankar, Saiprasad
    Zhou, Jiantao
    Zhu, Ce
    IEEE SIGNAL PROCESSING MAGAZINE, 2023, 40 (01) : 32 - 44