Self-aware cycle curriculum learning for multiple-choice reading comprehension

被引:0
作者
Chen, Haihong [1 ]
Li, Yufei [1 ]
机构
[1] Chifeng Univ, Sch Math & Comp Sci, Chifeng, Peoples R China
关键词
Machine reading comprehension; Multiple-choice; Self-aware; Cycle training strategy; Curriculum learning;
D O I
10.7717/peerj-cs.1179
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multiple-choice reading comprehension task has recently attracted significant interest. The task provides several options for each question and requires the machine to select one of them as the correct answer. Current approaches normally leverage a pre-training and then fine-tuning procedure that treats data equally, ignoring the difficulty of training examples. To solve this issue, curriculum learning (CL) has shown its effectiveness in improving the performance of models. However, previous methods have two problems with curriculum learning. First, most methods are rule-based, not flexible enough, and usually suitable for specific tasks, such as machine translation. Second, these methods arrange data from easy to hard or from hard to easy and overlook the fact that human beings usually learn from easy to difficult, and from difficult to easy when they make comprehension reading tasks. In this article, we propose a novel Self-Aware Cycle Curriculum Learning (SACCL) approach which can evaluate data difficulty from the model's perspective and train the model with cycle training strategy. The experiments show that the proposed approach achieves better performance on the C3 dataset than the baseline, which verifies the effectiveness of SACCL.
引用
收藏
页数:18
相关论文
共 41 条
[1]  
Bengio Y., 2009, P 26 ANN INT C MACH, DOI DOI 10.1145/1553374.1553380
[2]  
Büyüktas B, 2021, EUR SIGNAL PR CONF, P650, DOI 10.23919/Eusipco47968.2020.9287639
[3]  
Chang E, 2021, 16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), P727
[4]   A Multiple-Choice Machine Reading Comprehension Model with Multi-Granularity Semantic Reasoning [J].
Dai, Yu ;
Fu, Yufan ;
Yang, Lei .
APPLIED SCIENCES-BASEL, 2021, 11 (17)
[5]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[6]  
Gan ZF, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P3487
[7]  
Hacohen G, 2019, PR MACH LEARN RES, V97
[8]  
Hermann KM, 2015, ADV NEUR IN, V28
[9]   CurricularFace: Adaptive Curriculum Learning Loss for Deep Face Recognition [J].
Huang, Yuge ;
Wang, Yuhan ;
Tai, Ying ;
Liu, Xiaoming ;
Shen, Pengcheng ;
Li, Shaoxin ;
Li, Jilin ;
Huang, Feiyue .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :5900-5909
[10]  
Huang YY, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P389