Learning to Generate Questions by Learning to Recover Answer-containing Sentences

被引:0
|
作者
Back, Seohyun [1 ,2 ]
Kedia, Akhil [1 ]
Chinthakindi, Sai Chetan [1 ]
Lee, Haejun [1 ]
Choo, Jaegul [2 ]
机构
[1] Samsung Res, Seoul, South Korea
[2] Korea Adv Inst Sci & Technol, Daejeon, South Korea
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To train a question answering model based on machine reading comprehension (MRC), significant effort is required to prepare annotated training data composed of questions and their answers from contexts. Recent research has focused on synthetically generating a question from a given context and an annotated (or generated) answer by training an additional generative model to augment the training data. In light of this research direction, we propose a novel pre-training approach that learns to generate contextually rich questions, by recovering answer-containing sentences. We evaluate our method against existing ones in terms of the quality of generated questions, and fine-tuned MRC model accuracy after training on the data synthetically generated by our method. We consistently improve the question generation capability of existing models such as T5 and UniLM, and achieve state-of-the-art results on MS MARCO and NewsQA, and comparable results to the state-of-the-art on SQuAD. Additionally, the data synthetically generated by our approach is beneficial for boosting up the downstream MRC accuracy across a wide range of datasets, such as SQuAD-v1.1, v2.0, KorQuAD and BioASQ, without any modification to the existing MRC models. Furthermore, our method shines especially when a limited amount of pre-training or downstream MRC data is given.
引用
收藏
页码:1516 / 1529
页数:14
相关论文
共 50 条
  • [31] Learning to summarize and answer questions about a virtual robot's past actions
    DeChant, Chad
    Akinola, Iretiayo
    Bauer, Daniel
    AUTONOMOUS ROBOTS, 2023, 47 (08) : 1103 - 1118
  • [32] Learning to Answer Complex Visual Questions from Multi-View Analysis
    Zhu, Minjun
    Weng, Yixuan
    He, Shizhu
    Liu, Kang
    Zhao, Jun
    CCKS 2022 - EVALUATION TRACK, 2022, 1711 : 154 - 162
  • [33] Learning to summarize and answer questions about a virtual robot’s past actions
    Chad DeChant
    Iretiayo Akinola
    Daniel Bauer
    Autonomous Robots, 2023, 47 : 1103 - 1118
  • [34] LEARNING TO RECOVER SPARSE SIGNALS
    Zhong, Sichen
    Zhao, Yue
    Chen, Jianshu
    2019 57TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2019, : 995 - 1000
  • [35] Learning to Generate with Memory
    Li, Chongxuan
    Zhu, Jun
    Zhang, Bo
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [36] Learning to answer Emails
    Kockelkorn, M
    Lüneburg, A
    Scheffer, T
    ADVANCES IN INTELLIGENT DATA ANALYSIS V, 2003, 2810 : 25 - 35
  • [37] Learning to comprehend complex sentences
    Fincham, JM
    McClelland, JL
    PROCEEDINGS OF THE NINETEENTH ANNUAL CONFERENCE OF THE COGNITIVE SCIENCE SOCIETY, 1997, : 917 - 917
  • [38] SPECIFICITY IN LEARNING AND MEMORIZATION OF SENTENCES
    DEGOUYS, J
    ANNEE PSYCHOLOGIQUE, 1975, 75 (01): : 87 - 96
  • [39] Learning the Visual Interpretation of Sentences
    Zitnick, C. Lawrence
    Parikh, Devi
    Vanderwende, Lucy
    2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 1681 - 1688
  • [40] Learning to Prove Theorems by Learning to Generate Theorems
    Wang, Mingzhe
    Deng, Jia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33