Bayesian Multi-Task Transfer Learning for Soft Prompt Tuning

被引:0
|
作者
Lee, Haeju [1 ]
Jeong, Minchan [1 ]
Yun, Se-Young [1 ]
Kim, Kee-Eung [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Kim Jaechul Grad Sch AI, Seoul, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Prompt tuning, in which prompts are optimized to adapt large-scale pre-trained language models to downstream tasks instead of fine-tuning the full model parameters, has been shown to be particularly effective when the prompts are trained in the multi-task transfer learning setting. These methods generally involve individually training prompts for each source task and then aggregating them to provide the initialization of the prompt for the target task. However, this approach critically ignores the fact that some of the source tasks could be negatively or positively interfering with each other. We argue that when we extract knowledge from source tasks via training source prompts, we need to consider this correlation among source tasks for better transfer to target tasks. To this end, we propose a Bayesian approach where we work with the posterior distribution of prompts across source tasks. We obtain representative source prompts corresponding to the samples from the posterior utilizing Stein Variational Gradient Descent, which are then aggregated to constitute the initial target prompt. We show extensive experimental results on the standard benchmark NLP tasks, where our Bayesian multi-task transfer learning approach outperforms the state-of-the-art methods in many settings. Furthermore, our approach requires no auxiliary models other than the prompt itself, achieving high degree of parameter-efficiency.(1)
引用
收藏
页码:4942 / 4958
页数:17
相关论文
共 50 条
  • [41] Multi-Task Curriculum Transfer Deep Learning of Clothing Attributes
    Dong, Qi
    Gong, Shaogang
    Zhu, Xiatian
    2017 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2017), 2017, : 520 - 529
  • [42] Provably Efficient Multi-Task Reinforcement Learning with Model Transfer
    Zhang, Chicheng
    Wang, Zhi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [43] Efficient Multi-Task Learning via Iterated Single-Task Transfer
    Zentner, K. R.
    Puri, Ujjwal
    Zhang, Yulun
    Julian, Ryan
    Sukhatme, Gaurav S.
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 10141 - 10146
  • [44] Multi-task transfer learning for biomedical machine reading comprehension
    Guo, Wenyang
    Du, Yongping
    Zhao, Yiliang
    Ren, Keyan
    INTERNATIONAL JOURNAL OF DATA MINING AND BIOINFORMATICS, 2020, 23 (03) : 234 - 250
  • [45] Multi-task Transfer Learning for Timescale Graphical Event Models
    Monvoisin, Mathilde
    Leray, Philippe
    SYMBOLIC AND QUANTITATIVE APPROACHES TO REASONING WITH UNCERTAINTY, ECSQARU 2019, 2019, 11726 : 313 - 323
  • [46] Multi-Task Learning with Knowledge Transfer for Facial Attribute Classification
    Fanhe, Xiaohui
    Guo, Jie
    Huang, Zheng
    Qiu, Weidong
    Zhang, Yuele
    2019 IEEE INTERNATIONAL CONFERENCE ON INDUSTRIAL TECHNOLOGY (ICIT), 2019, : 877 - 882
  • [47] Match-Prompt: Improving Multi-task Generalization Ability for Neural Text Matching via Prompt Learning
    Xu, Shicheng
    Pang, Liang
    Shen, Huawei
    Cheng, Xueqi
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 2290 - 2300
  • [48] A Multi-Task and Transfer Learning based Approach for MOS Prediction
    Tian, Xiaohai
    Fu, Kaiqi
    Gao, Shaojun
    Gu, Yiwei
    Wang, Kai
    Li, Wei
    Ma, Zejun
    INTERSPEECH 2022, 2022, : 5438 - 5442
  • [49] Multi-population genomic prediction using a multi-task Bayesian learning model
    Chen, Liuhong
    Li, Changxi
    Miller, Stephen
    Schenkel, Flavio
    BMC GENETICS, 2014, 15
  • [50] Zero-Shot Rumor Detection via Meta Multi-Task Prompt Learning
    Shi, Yu
    Yu, Ning
    Sun, Yawei
    Liu, Jianyi
    Beijing Youdian Daxue Xuebao/Journal of Beijing University of Posts and Telecommunications, 2024, 47 (04): : 77 - 82