y-Tuning: an efficient tuning paradigm for large-scale pre-trained models via label representation learning

被引:0
|
作者
Liu, Yitao [1 ]
An, Chenxin [1 ]
Qiu, Xipeng [1 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai 200433, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
pre-trained model; lightweight fine-tuning paradigms; label representation;
D O I
10.1007/s11704-023-3131-8
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With current success of large-scale pre-trained models (PTMs), how efficiently adapting PTMs to downstream tasks has attracted tremendous attention, especially for PTMs with billions of parameters. Previous work focuses on designing parameter-efficient tuning paradigms but needs to save and compute the gradient of the whole computational graph. In this paper, we propose y-Tuning, an efficient yet effective paradigm to adapt frozen large-scale PTMs to specific downstream tasks. y-Tuning learns dense representations for labels y defined in a given task and aligns them to fixed feature representation. Without computing the gradients of text encoder at training phrase, y-Tuning is not only parameter-efficient but also training-efficient. Experimental results show that for DeBERTa(XXL) with 1.6 billion parameters, y-Tuning achieves performance more than 96% of full fine-tuning on GLUE Benchmark with only 2% tunable parameters and much fewer training costs.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] FedITD: A Federated Parameter-Efficient Tuning With Pre-Trained Large Language Models and Transfer Learning Framework for Insider Threat Detection
    Wang, Zhi Qiang
    Wang, Haopeng
    El Saddik, Abdulmotaleb
    IEEE ACCESS, 2024, 12 : 160396 - 160417
  • [22] 3D Semantic Novelty Detection via Large-Scale Pre-Trained Models
    Rabino, Paolo
    Alliegro, Antonio
    Tommasi, Tatiana
    IEEE ACCESS, 2024, 12 : 135352 - 135361
  • [23] Gender-tuning: Empowering Fine-tuning for Debiasing Pre-trained Language Models
    Ghanbarzadeh, Somayeh
    Huang, Yan
    Palangi, Hamid
    Moreno, Radames Cruz
    Khanpour, Hamed
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5448 - 5458
  • [24] Representation Transfer Learning via Multiple Pre-Trained Models for Linear Regression
    Singh, Navjot
    Diggavi, Suhas
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2025, 19 (01) : 208 - 220
  • [25] Attention Prompt Tuning: Parameter-efficient Adaptation of Pre-trained Models for Action Recognition
    Bandara, Wele Gedara Chaminda
    Patel, Vishal M.
    2024 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, FG 2024, 2024,
  • [26] Efficient Fine-Tuning for Low-Resource Tibetan Pre-trained Language Models
    Zhou, Mingjun
    Daiqing, Zhuoma
    Qun, Nuo
    Nyima, Tashi
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT VII, 2024, 15022 : 410 - 422
  • [27] DVPT: Dynamic Visual Prompt Tuning of large pre-trained models for medical image analysis
    He, Along
    Wu, Yanlin
    Wang, Zhihong
    Li, Tao
    Fu, Huazhu
    NEURAL NETWORKS, 2025, 185
  • [28] Towards Efficient Fine-Tuning of Pre-trained Code Models: An Experimental Study and Beyond
    Shi, Ensheng
    Wang, Yanlin
    Zhang, Hongyu
    Du, Lun
    Han, Shi
    Zhang, Dongmei
    Sun, Hongbin
    PROCEEDINGS OF THE 32ND ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2023, 2023, : 39 - 51
  • [29] Confounder balancing in adversarial domain adaptation for pre-trained large models fine-tuning
    Jiang, Shuoran
    Chen, Qingcai
    Xiang, Yang
    Pan, Youcheng
    Wu, Xiangping
    Lin, Yukang
    NEURAL NETWORKS, 2024, 173
  • [30] Fine-Tuning Pre-Trained Language Models with Gaze Supervision
    Deng, Shuwen
    Prasse, Paul
    Reich, David R.
    Scheffer, Tobias
    Jager, Lena A.
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 217 - 224