PELA: Learning Parameter-Efficient Models with Low-Rank Approximation

被引:0
|
作者
Guo, Yangyang [1 ]
Wang, Guangzhi [1 ]
Kankanhalli, Mohan [1 ]
机构
[1] Natl Univ Singapore, Singapore, Singapore
来源
2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2024年
基金
新加坡国家研究基金会;
关键词
D O I
10.1109/CVPR52733.2024.01486
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Applying a pre-trained large model to downstream tasks is prohibitive under resource-constrained conditions. Re-cent dominant approaches for addressing efficiency issues involve adding a few learnable parameters to the fixed backbone model. This strategy, however, leads to more challenges in loading large models for downstream fine-tuning with limited resources. In this paper, we propose a novel method for increasing the parameter efficiency of pre-trained models by introducing an intermediate pre-training stage. To this end, we first employ low-rank approximation to compress the original large model and then devise a feature distillation module and a weight perturbation regularization module. These modules are specifically designed to enhance the low-rank model. In particular, we update only the low-rank model while freezing the backbone parameters during pre-training. This allows for direct and efficient utilization of the low-rank model for downstream fine-tuning tasks. The proposed method achieves both efficiencies in terms of required parameters and computation time while maintaining comparable results with minimal modifications to the backbone architecture. Specifically, when applied to three vision-only and one vision-language Transformer models, our approach often demonstrates a merely similar to 0.6 point decrease in performance while reducing the original parameter size by 1/3 to 2/3. We release our code at link.
引用
收藏
页码:15699 / 15709
页数:11
相关论文
共 50 条
  • [41] Unsupervised feature selection via low-rank approximation and structure learning
    Wang, Shiping
    Wang, Han
    KNOWLEDGE-BASED SYSTEMS, 2017, 124 : 70 - 79
  • [42] On Approximation Algorithm for Orthogonal Low-Rank Tensor Approximation
    Yuning Yang
    Journal of Optimization Theory and Applications, 2022, 194 : 821 - 851
  • [43] Label Distribution Learning with Label Correlations via Low-Rank Approximation
    Ren, Tingting
    Jia, Xiuyi
    Li, Weiwei
    Zhao, Shu
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3325 - 3331
  • [44] On Approximation Algorithm for Orthogonal Low-Rank Tensor Approximation
    Yang, Yuning
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2022, 194 (03) : 821 - 851
  • [45] Accelerating Matrix Multiplication in Deep Learning by Using Low-Rank Approximation
    Osawa, Kazuki
    Sekiya, Akira
    Naganuma, Hiroki
    Yokota, Rio
    2017 INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING & SIMULATION (HPCS), 2017, : 186 - 192
  • [46] LEARNING REPRESENTATION FROM NEURAL FISHER KERNEL WITH LOW-RANK APPROXIMATION
    Zhang, Ruixiang
    Zhai, Shuangfei
    Littwin, Etai
    Susskind, Josh
    ICLR 2022 - 10th International Conference on Learning Representations, 2022,
  • [47] Low-Rank Approximation of Structural Redundancy for Self-Supervised Learning
    Du, Kang
    Xiang, Yu
    CAUSAL LEARNING AND REASONING, VOL 236, 2024, 236 : 1008 - 1032
  • [48] INFERENCE FOR LOW-RANK MODELS
    Chernozhukov, Victor
    Hansen, Christian
    Liao, Yuan
    Zhu, Yinchu
    ANNALS OF STATISTICS, 2023, 51 (03): : 1309 - 1330
  • [49] Efficient Stochastic Optimization for Low-Rank Distance Metric Learning
    Zhang, Jie
    Zhang, Lijun
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 933 - 939
  • [50] Parameter-Efficient Transfer Learning with Diff Pruning
    Guo, Demi
    Rush, Alexander M.
    Kim, Yoon
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4884 - 4896