SPAKT: A Self-Supervised Pre-TrAining Method for Knowledge Tracing

被引:5
作者
Ma, Yuling [1 ]
Han, Peng [2 ]
Qiao, Huiyan [1 ]
Cui, Chaoran [3 ]
Yin, Yilong [2 ]
Yu, Dehu [1 ]
机构
[1] Shandong Jianzhu Univ, Sch Comp Sci & Technol, Jinan 250101, Peoples R China
[2] Shandong Univ, Sch Software, Jinan 250101, Peoples R China
[3] Shandong Univ Finance & Econ, Sch Comp Sci & Technol, Jinan 250014, Peoples R China
基金
中国国家自然科学基金;
关键词
Knowledge engineering; Bayes methods; Logistics; Hidden Markov models; Education; Task analysis; Self-supervised learning; Knowledge tracing; student performance prediction; self-supervised learning; bidirectional encoder representation from transformers (BERT); BAYESIAN NETWORKS;
D O I
10.1109/ACCESS.2022.3187987
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Knowledge tracing (KT) is the core task of computer-aided education systems, and it aims at predicting whether a student can answer the next exercise (i.e., question) correctly based on his/her historical answer records. In recent years, deep neural network-based approaches have been widely developed in KT and achieved promising results. More recently, several researches further boost these KT models via exploiting plentiful relationships including exercise-skill relations (E-S), the exercise similarity (E-E) as well as skill similarity (S-S). However, these relationship information are frequently absent in many real-world educational applications, and it is a labor-intensive work for human experts to label it. Inspired by recent advances in natural language processing domain, we propose a novel pre-training approach, namely as SPAKT, and utilize self-supervised learning to pre-train exercise embedding representation without the need for expensive human-expert annotations in this paper. Contrary to existing pre-training methods that highly rely on manually labeling knowledge about the E-E, S-S, or E-S relationships, the core idea of the proposed SPAKT is to design three self-attention modules to model the E-S, E-E, and S-S relationships, respectively, and all of these three modules can be trained in the self-supervised setting. As a pre-training approach, our SPAKT can be effortlessly incorporated into existing deep neural network-based KT frameworks. We experimentally show that, even without using expensive annotations about the aforementioned three kinds of relationships, our model achieves competitive performance compared with state-of-the-arts. Our algorithm implementations have been made publicly available at https://github.com/Vinci-hp/pretrainKT.
引用
收藏
页码:72145 / 72154
页数:10
相关论文
共 35 条
[1]  
Baker RSJD, 2008, LECT NOTES COMPUT SC, V5091, P406
[2]  
Cen H, 2006, LECT NOTES COMPUT SC, V4053, P164
[3]   Tracking Knowledge Proficiency of Students with Educational Priors [J].
Chen, Yuying ;
Liu, Qi ;
Huang, Zhenya ;
Wu, Le ;
Chen, Enhong ;
Wu, Runze ;
Su, Yu ;
Hu, Guoping .
CIKM'17: PROCEEDINGS OF THE 2017 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2017, :989-998
[4]  
CORBETT AT, 1994, USER MODEL USER-ADAP, V4, P253, DOI 10.1007/BF01099821
[5]  
Devlin J, 2019, Arxiv, DOI [arXiv:1810.04805, 10.48550/arXiv.1810.04805]
[6]   Modeling learner's dynamic knowledge construction procedure and cognitive item difficulty for knowledge tracing [J].
Gan, Wenbin ;
Sun, Yuan ;
Peng, Xian ;
Sun, Yi .
APPLIED INTELLIGENCE, 2020, 50 (11) :3894-3912
[7]   Context-Aware Attentive Knowledge Tracing [J].
Ghosh, Aritra ;
Heffernan, Neil ;
Lan, Andrew S. .
KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, :2330-2339
[8]  
Gong Y., 2011, International Journal of Artificial Intelligence in Education, V21, P27
[9]  
Gonzalez-Brenes J., 2014, The 7th International Conference on Educational Data Mining, P84
[10]   AutoTutor: An intelligent tutoring system with mixed-initiative dialogue [J].
Graesser, AC ;
Chipman, P ;
Haynes, BC ;
Olney, A .
IEEE TRANSACTIONS ON EDUCATION, 2005, 48 (04) :612-618