Modelling the Process of Learning Analytics Using a Reinforcement Learning Framework

被引:1
作者
Choi, Samuel P. M. [1 ]
Lam, Franklin S. S. [1 ]
机构
[1] Open Univ Hong Kong, Lee Shau Kee Sch Business & Adm, Hong Kong, Hong Kong, Peoples R China
来源
INNOVATIONS IN OPEN AND FLEXIBLE EDUCATION | 2018年
关键词
Learning analytics; Learning analytics process; Reinforcement learning; Machine learning; Optimisation;
D O I
10.1007/978-981-10-7995-5_22
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Learning analytics (LA) is a relatively new research field concerned with analysing data collected from various sources to provide insights into enhancing learning and teaching. A complete LA process typically involves five distinct, yet interrelated, stages - namely capture, report, predict, act and refine - which form a sequential decision process. So far, research efforts have focused mostly on studying independent research questions involved in individual stages. It is therefore necessary to have a formal framework to quantify and guide the whole LA process. In this paper, we discuss how reinforcement learning (RL), a subfield of machine learning, can be employed to address the sequential decision problem involved in the LA process. In particular, we integrate the LA stages with an RL framework consisting of state space, action space, transition function and reward function and illustrate this with examples of how the three most studied optimality criteria in RL - finite horizon, discounted infinite horizon and the average reward model - can be applied to the LA process. The underlying assumptions, advantages and issues in the proposed RL framework are also discussed.
引用
收藏
页码:243 / 251
页数:9
相关论文
共 8 条
[1]  
[Anonymous], 2012, P 2 INT C LEARN AN K, DOI [10.1145/2330601.2330666, DOI 10.1145/2330601.2330666]
[2]  
[Anonymous], LAK12
[3]  
Campbell J., 2007, Educause Quarterly
[4]   Classroom walls that talk: Using online course activity data of successful students to raise self-awareness of underperforming peers [J].
Fritz, John .
INTERNET AND HIGHER EDUCATION, 2011, 14 (02) :89-97
[5]  
Long P., 2011, P 1 INT C LEARN AN K
[6]  
Puterman ML, 1994, MARKOV DECISION PROC
[7]  
Spaan MTJ, 2012, ADAPT LEARN OPTIM, V12, P387
[8]  
van Otterlo M, 2012, ADAPT LEARN OPTIM, V12, P3