Qadg: Generating question–answer-distractors pairs for real examination

被引:0
作者
Zhou, Hao [1 ]
Li, Li [1 ]
机构
[1] School of Computer & Information Science, Southwest University, Chongqing, Chongqing
基金
中国国家自然科学基金;
关键词
Distractor generation; Natural language processing; Pre-trained model; Question generation;
D O I
10.1007/s00521-024-10658-5
中图分类号
学科分类号
摘要
Reading comprehension question generation aims to generate questions from a given article, while distractor generation involves generating multiple distractors from a given article, question, and answer. Most existing research has mainly focused on one of the above tasks, with limited attention to the joint task of Question–Answer-Distractor (QAD) generation. While previous work has achieved success in the joint generation of answer-aware questions and distractors, applying these answer-aware approaches to practical applications in the education domain remains challenging. In this study, we propose a unified and high-performance Question–Answer-Distractors Generation model, named QADG. Our model comprises two components: Question–Answer Generation (QAG) and Distractor Generation (DG). This model is capable of generating Question–Answer pairs based on a given context and then generating distractors based on the context and QA pairs. To address the unconstrained nature of question-and-answer generation in QAG, we employ a key phrase extraction as reported by Willis (in: proceedings of the Sixth ACM Conference on Learning@ Scale, 2019) module to extract key phrases from the article. The extracted key phrases, as the constraints that can be used to match answers. To enhance the quality of distractors, we propose a novel ranking-rewriting mechanism. We employ a fine-tuned model to rank distractors and introduce a rewriting module to improve the quality of distractors. Furthermore, the Knowledge-Dependent-Answerability (KDA) as reported by Moon (Evaluating the knowledge dependency of questions, 2022) is used as a filter to ensure the answerability of the generated QAD pairs. Experiments on SQuAD and RACE datasets demonstrate that the proposed QADG exhibits superior performance, particularly in the DG phase. Additionally, human evaluations also confirm the effectiveness and educational relevance of our model. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2024.
引用
收藏
页码:1157 / 1170
页数:13
相关论文
共 57 条
[1]  
Willis A., Davis G., Ruan S., Manoharan L., Landay J., Brunskill E., Key phrase extraction for generating educational question-answer pairs, Proceedings of the Sixth ACM Conference on Learning@ Scale, pp. 1-10, (2019)
[2]  
Moon H., Yang Y., Shin J., Yu H., Lee S., Jeong M., Park J., Kim M., Choi S., Evaluating the knowledge dependency of questions, . Arxiv Preprint Arxiv, 2211, (2022)
[3]  
Lai G., Xie Q., Liu H., Yang Y., Hovy E., Race Large-Scale Reading Comprehension Dataset from Examinations, (2017)
[4]  
Zhou Q., Yang N., Wei F., Tan C., Bao H., Zhou M., Neural Question Generation from Text: A Preliminary Study, 6, pp. 662-671, (2018)
[5]  
Zhao Y., Ni X., Ding Y., Ke Q., Paragraph-level neural question generation with maxout pointer and gated self-attention networks. In: proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp, 3901–3910, (2018)
[6]  
Qi W., Yan Y., Gong Y., Liu D., Duan N., Chen J., Zhang R., Zhou M., Prophetnet: Predicting future n-gram for sequence-to-sequence pre-training, . Arxiv Preprint Arxiv, 2001, (2020)
[7]  
Jia X., Zhou W., Sun X., Wu Y., How to ask good questions? Try to leverage paraphrases, Proceedings of the 58Th Annual Meeting of the Association for Computational Linguistics, pp. 6130-6140, (2020)
[8]  
Sun Y., Liu S., Dan Z., Zhao X., Question generation based on grammar knowledge and fine-grained classification, Proceedings of the 29Th International Conference on Computational Linguistics, pp. 6457-6467, (2022)
[9]  
Wang S., Wei Z., Fan Z., Liu Y., Huang X., A multi-agent communication framework for question-worthy phrase extraction and question generation, proceedings of the AAAI Conference on Artificial Intelligence, 33, pp. 7168-7175, (2019)
[10]  
Cui S., Bao X., Zu X., Guo Y., Zhao Z., Zhang J., Chen H., Onestop qamaker: Extract question-answer pairs from text in a one-stop approach., (2021)