Meta-training with Demonstration Retrieval for Efficient Few-shot Learning

被引:0
|
作者
Mueller, Aaron [1 ]
Narang, Kanika [2 ]
Mathias, Lambert [2 ]
Wang, Qifan [2 ]
Firooz, Hamed [2 ]
机构
[1] Johns Hopkins Univ, Baltimore, MD 21218 USA
[2] Meta AI, Menlo Pk, CA USA
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models show impressive results on few-shot NLP tasks. However, these models are memory and computation-intensive. Meta-training allows one to leverage smaller models for few-shot generalization in a domain-general and task-agnostic manner (Min et al., 2022a; Wei et al., 2022; Chen et al., 2022); however, these methods alone results in models that may not have sufficient parameterization or knowledge to adapt quickly to a large variety of tasks. To overcome this issue, we propose meta-training with demonstration retrieval, where we use a dense passage retriever to retrieve semantically similar labeled demonstrations to each example for more varied supervision. By separating external knowledge from model parameters, we can use meta-training to train parameter-efficient models that generalize well on a larger variety of tasks. We construct a meta-training set from UNIFIEDQA and CROSSFIT, and propose a demonstration bank based on UNIFIEDQA tasks. To our knowledge, our work is the first to combine retrieval with meta-training, to use DPR models to retrieve demonstrations, and to leverage demonstrations from many tasks simultaneously, rather than randomly sampling demonstrations from the training set of the target task. Our approach outperforms a variety of targeted parameter-efficient and retrieval-augmented few-shot methods on QA, NLI, and text classification tasks (including SQuAD, QNLI, and TREC). Our approach can be metatrained and fine-tuned quickly on a single GPU.
引用
收藏
页码:6049 / 6064
页数:16
相关论文
共 50 条
  • [21] Demonstration-Conditioned Reinforcement Learning for Few-Shot Imitation
    Cachet, Theo
    Perez, Julien
    Dance, Christopher R.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [22] META-LEARNING WITH ATTENTION FOR IMPROVED FEW-SHOT LEARNING
    Hou, Zejiang
    Walid, Anwar
    Kung, Sun-Yuan
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2725 - 2729
  • [23] Meta-pruning: Learning to Prune on Few-Shot Learning
    Chu, Yan
    Liu, Keshi
    Jiang, Songhao
    Sun, Xianghui
    Wang, Baoxu
    Wang, Zhengkui
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I, KSEM 2024, 2024, 14884 : 74 - 85
  • [24] Task Agnostic Meta-Learning for Few-Shot Learning
    Jamal, Muhammad Abdullah
    Qi, Guo-Jun
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 11711 - 11719
  • [25] Few-shot Metric Learning: Online Adaptation of Embedding for Retrieval
    Jung, Deunsol
    Kang, Dahyun
    Kwak, Suha
    Cho, Minsu
    COMPUTER VISION - ACCV 2022, PT V, 2023, 13845 : 54 - 70
  • [26] FEW-SHOT LEARNING FOR REMOTE SENSING IMAGE RETRIEVAL WITH MAML
    Zhong, Qian
    Chen, Ling
    Qian, Yuntao
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 2446 - 2450
  • [27] Few-Shot Composition Learning for Image Retrieval with Prompt Tuning
    Wu, Junda
    Wang, Rui
    Zhao, Handong
    Zhang, Ruiyi
    Lu, Chaochao
    Li, Shuai
    Henao, Ricardo
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 4, 2023, : 4729 - 4737
  • [28] Fair Meta-Learning For Few-Shot Classification
    Zhao, Chen
    Li, Changbin
    Li, Jincheng
    Chen, Feng
    11TH IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH (ICKG 2020), 2020, : 275 - 282
  • [29] Meta-BN Net for few-shot learning
    GAO Wei
    SHAO Mingwen
    SHU Jun
    ZHUANG Xinkai
    Frontiers of Computer Science, 2023, 17 (01)
  • [30] Meta-BN Net for few-shot learning
    Wei Gao
    Mingwen Shao
    Jun Shu
    Xinkai Zhuang
    Frontiers of Computer Science, 2023, 17