CONTEXTUAL MULTI-ARMED BANDIT ALGORITHMS FOR PERSONALIZED LEARNING ACTION SELECTION

被引:0
|
作者
Manickam, Indu [1 ]
Lan, Andrew S. [1 ]
Baraniuk, Richard G. [1 ]
机构
[1] Rice Univ, Houston, TX 77251 USA
关键词
contextual bandits; personalized learning;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Optimizing the selection of learning resources and practice questions to address each individual student's needs has the potential to improve students' learning efficiency. In this paper, we study the problem of selecting a personalized learning action for each student (e. g. watching a lecture video, working on a practice question, etc.), based on their prior performance, in order to maximize their learning outcome. We formulate this problem using the contextual multi-armed bandits framework, where students' prior concept knowledge states (estimated from their responses to questions in previous assessments) correspond to contexts, the personalized learning actions correspond to arms, and their performance on future assessments correspond to rewards. We propose three new Bayesian policies to select personalized learning actions for students that each exhibits advantages over prior work, and experimentally validate them using real-world datasets.
引用
收藏
页码:6344 / 6348
页数:5
相关论文
共 50 条
  • [41] Transfer Learning in Multi-Armed Bandit: A Causal Approach
    Zhang, Junzhe
    Bareinboim, Elias
    AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2017, : 1778 - 1780
  • [42] Achieving Complete Learning in Multi-Armed Bandit Problems
    Vakili, Sattar
    Zhao, Qing
    2013 ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS AND COMPUTERS, 2013, : 1778 - 1782
  • [43] Distributed Learning in Multi-Armed Bandit With Multiple Players
    Liu, Keqin
    Zhao, Qing
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2010, 58 (11) : 5667 - 5681
  • [44] Contextual Multi-Armed Bandit-Based Link Adaptation for URLLC
    Ku, Sungmo
    Lee, Chungyong
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (11) : 17305 - 17315
  • [45] Reinforcement learning and evolutionary algorithms for non-stationary multi-armed bandit problems
    Koulouriotis, D. E.
    Xanthopoulos, A.
    APPLIED MATHEMATICS AND COMPUTATION, 2008, 196 (02) : 913 - 922
  • [46] Multi-armed bandit algorithms over DASH for multihomed client
    Hodroj, Ali
    Ibrahim, Marc
    Hadjadj-Aoul, Yassine
    Sericola, Bruno
    INTERNATIONAL JOURNAL OF SENSOR NETWORKS, 2021, 37 (04) : 244 - 253
  • [47] Reconfigurable and Computationally Efficient Architecture for Multi-armed Bandit Algorithms
    Santosh, S. V. Sai
    Darak, S. J.
    2020 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2020,
  • [48] A Hybrid Proactive Caching System in Vehicular Networks Based on Contextual Multi-Armed Bandit Learning
    Wang, Qiao
    Grace, David
    IEEE ACCESS, 2023, 11 : 29074 - 29090
  • [49] CHANNEL SELECTION WITH RAYLEIGH FADING: A MULTI-ARMED BANDIT FRAMEWORK
    Jouini, Wassim
    Moy, Christophe
    2012 IEEE 13TH INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (SPAWC), 2012, : 299 - 303
  • [50] muMAB: A Multi-Armed Bandit Model for Wireless Network Selection
    Boldrini, Stefano
    De Nardis, Luca
    Caso, Giuseppe
    Le, Mai T. P.
    Fiorina, Jocelyn
    Di Benedetto, Maria-Gabriella
    ALGORITHMS, 2018, 11 (02)