Deep Knowledge Tracing on Skills with Small Datasets

被引:1
作者
Tato, Ange [1 ]
Nkambou, Roger [1 ]
机构
[1] Univ Quebec Montreal, Montreal, PQ, Canada
来源
INTELLIGENT TUTORING SYSTEMS, ITS 2022 | 2022年 / 13284卷
关键词
Deep Knowledge Tracing; Bayesian Knowledge Tracing; Knowledge Tracing;
D O I
10.1007/978-3-031-09680-8_12
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Knowledge Tracing (DKT), as well as other machine learning approaches, is biased toward data used during the training step. Thus, for problems where we have few amounts of data for training, the generalization power will be low, and the models will tend to work well on classes containing many samples and poorly on those with few. This situation is quite common in educational data where some skills are very difficult to master while others are very easy. As a result, there will be less data on students who correctly answered questions related to difficult skills, but also on those who provided incorrect answers to questions related to easy skills. In those cases, the DKT is unable to correctly predict the student's answers to questions associated with these skills. To improve DKT performance under these conditions, we have developed a two-fold approach. Firstly, the loss function is modified so that some skills are masked to force the model's attention on those that are difficult to generalize. Secondly, to cope with the limited amount of data on some skills, we proposed a hybrid architecture that integrates a priori (expert) knowledge with DKT through an attentional mechanism. The resulting model accurately tracks student Knowledge in the Logic-Muse Intelligent Tutoring System (ITS), compared to the traditional Bayesian Knowledge Tracing (BKT) and the original DKT.
引用
收藏
页码:123 / 135
页数:13
相关论文
共 31 条
[1]  
Ange T., 2018, 2018 INT JOINT C NEU, P1
[2]  
[Anonymous], 2008, P 25 INT C MACH LEAR, DOI DOI 10.1145/1390156.1390177
[3]  
Bahdanau D, 2016, Arxiv, DOI [arXiv:1409.0473, DOI 10.48550/ARXIV.1409.0473]
[4]  
Beck JE, 2007, LECT NOTES ARTIF INT, V4511, P137
[5]  
Chorowski J, 2015, ADV NEUR IN, V28
[6]  
CORBETT AT, 1994, USER MODEL USER-ADAP, V4, P253, DOI 10.1007/BF01099821
[7]  
Desmarais M., 2019, P 12 INT C ED DAT MI, P623
[8]  
Graves A., 2013, Generating sequences with recurrent neural networks
[9]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[10]   Learning Deep Representation for Imbalanced Classification [J].
Huang, Chen ;
Li, Yining ;
Loy, Chen Change ;
Tang, Xiaoou .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :5375-5384