Bayesian Multi-Task Transfer Learning for Soft Prompt Tuning

被引:0
|
作者
Lee, Haeju [1 ]
Jeong, Minchan [1 ]
Yun, Se-Young [1 ]
Kim, Kee-Eung [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Kim Jaechul Grad Sch AI, Seoul, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prompt tuning, in which prompts are optimized to adapt large-scale pre-trained language models to downstream tasks instead of fine-tuning the full model parameters, has been shown to be particularly effective when the prompts are trained in the multi-task transfer learning setting. These methods generally involve individually training prompts for each source task and then aggregating them to provide the initialization of the prompt for the target task. However, this approach critically ignores the fact that some of the source tasks could be negatively or positively interfering with each other. We argue that when we extract knowledge from source tasks via training source prompts, we need to consider this correlation among source tasks for better transfer to target tasks. To this end, we propose a Bayesian approach where we work with the posterior distribution of prompts across source tasks. We obtain representative source prompts corresponding to the samples from the posterior utilizing Stein Variational Gradient Descent, which are then aggregated to constitute the initial target prompt. We show extensive experimental results on the standard benchmark NLP tasks, where our Bayesian multi-task transfer learning approach outperforms the state-of-the-art methods in many settings. Furthermore, our approach requires no auxiliary models other than the prompt itself, achieving high degree of parameter-efficiency.(1)
引用
收藏
页码:4942 / 4958
页数:17
相关论文
共 50 条
  • [21] Driver Drowsiness Detection by Multi-task and Transfer Learning
    Chang, Yuan
    Kameyama, Wataru
    INTERNATIONAL WORKSHOP ON ADVANCED IMAGING TECHNOLOGY (IWAIT) 2022, 2022, 12177
  • [22] Episodic memory transfer for multi-task reinforcement learning
    Sorokin, Artyom Y.
    Burtsev, Mikhail S.
    BIOLOGICALLY INSPIRED COGNITIVE ARCHITECTURES, 2018, 26 : 91 - 95
  • [23] Multi-task Pre-training with Soft Biometrics for Transfer-learning Palmprint Recognition
    Xu, Huanhuan
    Leng, Lu
    Yang, Ziyuan
    Teoh, Andrew Beng Jin
    Jin, Zhe
    NEURAL PROCESSING LETTERS, 2023, 55 (03) : 2341 - 2358
  • [24] Multi-task Pre-training with Soft Biometrics for Transfer-learning Palmprint Recognition
    Huanhuan Xu
    Lu Leng
    Ziyuan Yang
    Andrew Beng Jin Teoh
    Zhe Jin
    Neural Processing Letters, 2023, 55 : 2341 - 2358
  • [25] Bayesian multi-task learning for decoding multi-subject neuroimaging data
    Marquand, Andre F.
    Brammer, Michael
    Williams, Steven C. R.
    Doyle, Orla M.
    NEUROIMAGE, 2014, 92 : 298 - 311
  • [26] Multi-Subject EEG Classification: Bayesian Nonparametrics and Multi-Task Learning
    Choi, Seungjin
    3RD INTERNATIONAL WINTER CONFERENCE ON BRAIN-COMPUTER INTERFACE, 2015, : 35 - 35
  • [27] Parameter Efficient Multi-task Fine-tuning by Learning to Transfer Token-wise Prompts
    Wu, Muling
    Liu, Wenhao
    Xu, Jianhan
    Lv, Changze
    Ling, ZiXuan
    Li, Tianlong
    Huang, LongTao
    Zheng, XiaoQing
    Huang, Xuanjing
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 8734 - 8746
  • [28] ON PARALLELIZING MULTI-TASK BAYESIAN OPTIMIZATION
    Groves, Matthew
    Pearce, Michael
    Branke, Juergen
    2018 WINTER SIMULATION CONFERENCE (WSC), 2018, : 1993 - 2002
  • [29] MmAP : Multi-Modal Alignment Prompt for Cross-Domain Multi-Task Learning
    Xin, Yi
    Du, Junlong
    Wang, Qiang
    Yan, Ke
    Ding, Shouhong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 14, 2024, : 16076 - 16084
  • [30] Bayesian Max-margin Multi-Task Learning with Data Augmentation
    Li, Chengtao
    Zhu, Jun
    Chen, Lianfei
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 415 - 423