Few-Shot Question Answering by Pretraining Span Selection

被引:0
|
作者
Ram, Ori [1 ]
Kirstain, Yuval [1 ]
Berant, Jonathan [1 ,2 ]
Globerson, Amir [1 ]
Levy, Omer [1 ]
机构
[1] Tel Aviv Univ, Blavatnik Sch Comp Sci, Tel Aviv, Israel
[2] Allen Inst AI, Seattle, WA USA
来源
59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1 | 2021年
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In several question answering benchmarks, pretrained models have reached human parity through fine-tuning on an order of 100,000 annotated questions and answers. We explore the more realistic few-shot setting, where only a few hundred training examples are available, and observe that standard models perform poorly, highlighting the discrepancy between current pretraining objectives and question answering. We propose a new pretraining scheme tailored for question answering: recurring span selection. Given a passage with multiple sets of recurring spans, we mask in each set all recurring spans but one, and ask the model to select the correct span in the passage for each masked span. Masked spans are replaced with a special token, viewed as a question representation, that is later used during fine-tuning to select the answer span. The resulting model obtains surprisingly good results on multiple benchmarks (e.g., 72.7 F1 on SQuAD with only 128 training examples), while maintaining competitive performance in the high-resource setting.(1)
引用
收藏
页码:3066 / 3079
页数:14
相关论文
共 50 条
  • [1] OmniTab: Pretraining with Natural and Synthetic Data for Few-shot Table-based Question Answering
    Jiang, Zhengbao
    Mao, Yi
    He, Pengcheng
    Neubig, Graham
    Chen, Weizhu
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 932 - 942
  • [2] QARR-FSQA: Question-Answer Replacement and Removal Pretraining Framework for Few-Shot Question Answering
    Tan, Siao Wah
    Lee, Chin Poo
    Lim, Kian Ming
    Tee, Connie
    Alqahtani, Ali
    IEEE ACCESS, 2024, 12 : 159280 - 159295
  • [3] Few-Shot Multihop Question Answering over Knowledge Base
    Fan, Meihao
    Zhang, Lei
    Xiao, Siyao
    Liang, Yuru
    WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2022, 2022
  • [4] Few-shot Unified Question Answering: Tuning Models or Prompts?
    Bansal, Srijan
    Yavuz, Semih
    Pang, Bo
    Bhat, Meghana
    Zhou, Yingbo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 8200 - 8220
  • [5] Explore pretraining for few-shot learning
    Li, Yan
    Huang, Jinjie
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (2) : 4691 - 4702
  • [6] Explore pretraining for few-shot learning
    Yan Li
    Jinjie Huang
    Multimedia Tools and Applications, 2024, 83 : 4691 - 4702
  • [7] Few-shot In-context Learning for Knowledge Base Question Answering
    Li, Tianle
    Ma, Xueguang
    Zhuang, Alex
    Gu, Yu
    Su, Yu
    Chen, Wenhu
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 6966 - 6980
  • [8] Domain-Specific Few-Shot Table Prompt Question Answering via Contrastive Exemplar Selection
    Mo, Tianjin
    Xiao, Qiao
    Zhang, Hongyi
    Li, Ren
    Wu, Yunsong
    ALGORITHMS, 2024, 17 (07)
  • [9] Zero-Shot and Few-Shot Video Question Answering with Multi-Modal Prompts
    Engin, Deniz
    Avrithis, Yannis
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 2796 - 2802
  • [10] Fast Parameter Adaptation for Few-shot Image Captioning and Visual Question Answering
    Dong, Xuanyi
    Zhu, Linchao
    Zhang, De
    Yang, Yi
    Wu, Fei
    PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 54 - 62