A Pre-trained Knowledge Tracing Model with Limited Data

被引:0
作者
Yue, Wenli [1 ,3 ]
Su, Wei [1 ,3 ]
Liu, Lei [2 ]
Cai, Chuan [1 ]
Yuan, Yongna [1 ]
Jia, Zhongfeng [1 ]
Liu, Jiamin [1 ]
Xie, Wenjian [1 ]
机构
[1] Lanzhou Univ, Sch Informat Sci & Engn, Lanzhou, Peoples R China
[2] Duzhe Publishing Grp Co Ltd, Lanzhou, Peoples R China
[3] Key Lab Media Convergence Technol & Commun, Lanzhou, Gansu, Peoples R China
来源
DATABASE AND EXPERT SYSTEMS APPLICATIONS, PT I, DEXA 2024 | 2024年 / 14910卷
关键词
Knowledge Tracing; Limited Data; Pre-training; Fine-tuning;
D O I
10.1007/978-3-031-68309-1_14
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Online education systems have gained increasing popularity due to their capability to fully preserve users' learning data. This advantage enables researchers to assess learners' mastery through their learning trajectories, thereby facilitating personalized education and support. Knowledge tracing, an effective educational aid, simulates students' implicit knowledge states and predicts their mastery over knowledge based on their historical answer records. However, for newly developed online learning platforms, the lack of sufficient historical answer data may impede accurate prediction of students' knowledge states, rendering existing knowledge tracing models less effective. This paper introduces the first pre-trained knowledge tracing model that leverages a substantial amount of existing data for pre-training and a smaller dataset for fine-tuning. Validated across several publicly available knowledge tracing datasets, our method demonstrates significant improvement in tracing performance on small datasets, with a maximum AUC increase of 5.07%. Beyond incorporating small datasets, our approach of pre-training the entire dataset has shown an enhanced AUC compared to the baseline, marking a novel direction in knowledge tracing research. Furthermore, the paper analyzed the outcomes of pre-training experiments with varying numbers of interactions as fine-tuning datasets, providing valuable insights for Intelligent Tutoring Systems (ITS).
引用
收藏
页码:163 / 178
页数:16
相关论文
共 50 条
[21]   CollRec: Pre-Trained Language Models and Knowledge Graphs Collaborate to Enhance Conversational Recommendation System [J].
Liu, Shuang ;
Ao, Zhizhuo ;
Chen, Peng ;
Kolmanic, Simon .
IEEE ACCESS, 2024, 12 :104663-104675
[22]   VTrans: A VAE-Based Pre-Trained Transformer Method for Microbiome Data Analysis [J].
Shi, Xinyuan ;
Zhu, Fangfang ;
Min, Wenwen .
JOURNAL OF COMPUTATIONAL BIOLOGY, 2025,
[23]   Data-Efficient Information Extraction from Documents with Pre-trained Language Models [J].
Sage, Clement ;
Douzon, Thibault ;
Aussem, Alex ;
Eglin, Veronique ;
Elghazel, Haytham ;
Duffner, Stefan ;
Garcia, Christophe ;
Espinas, Jeremy .
DOCUMENT ANALYSIS AND RECOGNITION, ICDAR 2021, PT II, 2021, 12917 :455-469
[24]   Novel Fine-Tuning Strategy on Pre-trained Protein Model Enhances ACP Functional Type Classification [J].
Wang, Shaokai ;
Ma, Bin .
BIOINFORMATICS RESEARCH AND APPLICATIONS, PT I, ISBRA 2024, 2024, 14954 :371-382
[25]   A Pre-Trained Deep Learning Model for Fast Online Prediction of Structural Seismic Responses [J].
Tang, Wei-Jian ;
Wang, Dong-Sheng ;
Huang, Hai-Bin ;
Dai, Jian-Cheng ;
Shi, Fan .
INTERNATIONAL JOURNAL OF STRUCTURAL STABILITY AND DYNAMICS, 2024, 24 (14)
[26]   SAILER: Structure-aware Pre-trained Language Model for Legal Case Retrieval [J].
Li, Haitao ;
Ai, Qingyao ;
Chen, Jia ;
Dong, Qian ;
Wu, Yueyue ;
Liu, Yiqun ;
Chen, Chong ;
Tian, Qi .
PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, :1035-1044
[27]   scMulan: A Multitask Generative Pre-Trained Language Model for Single-Cell Analysis [J].
Bian, Haiyang ;
Chen, Yixin ;
Dong, Xiaomin ;
Li, Chen ;
Hao, Minsheng ;
Chen, Sijie ;
Hu, Jinyi ;
Sun, Maosong ;
Wei, Lei ;
Zhang, Xuegong .
RESEARCH IN COMPUTATIONAL MOLECULAR BIOLOGY, RECOMB 2024, 2024, 14758 :479-482
[28]   Bird Whisperer: Leveraging Large Pre-trained Acoustic Model for Bird Call Classification [J].
Sheikh, Muhammad Umer ;
Abid, Hassan ;
Shafique, Bhuiyan Sanjid ;
Hanif, Asif ;
Haris, Muhammad .
INTERSPEECH 2024, 2024, :5028-5032
[29]   Comparative Analysis of Pre-trained Deep Neural Networks for Plant Disease Classification [J].
George, Romiyal ;
Thuseethan, Selvarajah ;
Ragel, Roshan G. .
2024 21ST INTERNATIONAL JOINT CONFERENCE ON COMPUTER SCIENCE AND SOFTWARE ENGINEERING, JCSSE 2024, 2024, :179-186
[30]   A Comparative Study of Using Pre-trained Language Models for Toxic Comment Classification [J].
Zhao, Zhixue ;
Zhang, Ziqi ;
Hopfgartner, Frank .
WEB CONFERENCE 2021: COMPANION OF THE WORLD WIDE WEB CONFERENCE (WWW 2021), 2021, :500-507