Modeling Extractive Question Answering Using Encoder-Decoder Models with Constrained Decoding and Evaluation-Based Reinforcement Learning

被引:2
作者
Li, Shaobo [1 ]
Sun, Chengjie [1 ]
Liu, Bingquan [1 ]
Liu, Yuanchao [1 ]
Ji, Zhenzhou [1 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Technol, Harbin 150001, Peoples R China
基金
中国国家自然科学基金;
关键词
natural language processing; question answering; encoder-decoder models; reinforcement learning; neural network; machine learning;
D O I
10.3390/math11071624
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Extractive Question Answering, also known as machine reading comprehension, can be used to evaluate how well a computer comprehends human language. It is a valuable topic with many applications, such as in chatbots and personal assistants. End-to-end neural-network-based models have achieved remarkable performance on these tasks. The most frequently used approach to extract answers with neural networks is to predict the answer's start and end positions in the document, independently or jointly. In this paper, we propose another approach that considers all words in an answer jointly. We introduce an encoder-decoder model to learn from all words in the answer. This differs from previous works. which usually focused on the start and end and ignored the words in the middle. To help the encoder-decoder model to perform this task better, we employ evaluation-based reinforcement learning with different reward functions. The results of an experiment on the SQuAD dataset show that the proposed method can outperform the baseline in terms of F1 scores, offering another potential approach to solve the extractive QA task.
引用
收藏
页数:16
相关论文
共 63 条
  • [1] A survey of longest common subsequence algorithms
    Bergroth, L
    Hakonen, H
    Raita, T
    [J]. SPIRE 2000: SEVENTH INTERNATIONAL SYMPOSIUM ON STRING PROCESSING AND INFORMATION RETRIEVAL - PROCEEDINGS, 2000, : 39 - 48
  • [2] Bowman Samuel R., 2015, P 2015 C EMPIRICAL M, DOI DOI 10.18653/V1/D15-1075
  • [3] Chen Danqi, 2018, Neural reading comprehension and beyond
  • [4] Chen YC, 2018, PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, P675
  • [5] Cho K, 2014, ARXIV14061078, V1, P1724, DOI [10.3115/v1/D14-1179, DOI 10.3115/V1/D14-1179]
  • [6] Clark K., 2020, ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS, DOI [10.48550/arXiv.2003.10555, DOI 10.48550/arXiv.2003.10555]
  • [7] Dahlmeier Daniel, 2012, HUMAN LANGUAGE TECHN, P568
  • [8] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
  • [9] Fajcik Martin, 2021, P 3 WORKSH MACH READ, P14
  • [10] Fan A, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P3558