Personalized Treatment Policies with the Novel Buckley-James Q-Learning Algorithm

被引:3
|
作者
Lee, Jeongjin [1 ]
Kim, Jong-Min [2 ]
机构
[1] Ohio State Univ, Dept Stat, Columbus, OH 43210 USA
[2] Univ Minnesota Morris, Stat Discipline, Morris, MN 56267 USA
关键词
Q-learning; reinforcement learning; precision medicine; Buckley-James Method; survival analysis; DYNAMIC TREATMENT REGIMES; LINEAR-REGRESSION; SURVIVAL;
D O I
10.3390/axioms13040212
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This research paper presents the Buckley-James Q-learning (BJ-Q) algorithm, a cutting-edge method designed to optimize personalized treatment strategies, especially in the presence of right censoring. We critically assess the algorithm's effectiveness in improving patient outcomes and its resilience across various scenarios. Central to our approach is the innovative use of the survival time to impute the reward in Q-learning, employing the Buckley-James method for enhanced accuracy and reliability. Our findings highlight the significant potential of personalized treatment regimens and introduce the BJ-Q learning algorithm as a viable and promising approach. This work marks a substantial advancement in our comprehension of treatment dynamics and offers valuable insights for augmenting patient care in the ever-evolving clinical landscape.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] An algorithm for the Buckley-James estimator with interval-censored data
    Yu, Qiqing
    Wong, George Y. C.
    JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2009, 79 (11) : 1341 - 1353
  • [2] Backward Q-learning: The combination of Sarsa algorithm and Q-learning
    Wang, Yin-Hao
    Li, Tzuu-Hseng S.
    Lin, Chih-Jui
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2013, 26 (09) : 2184 - 2193
  • [3] Buckley-James boosting model based on extreme learning machine and random survival forests
    Kong, Jianfen
    Zhang, Shuhong
    BIOMETRICAL JOURNAL, 2023, 65 (05)
  • [4] ENHANCEMENTS OF FUZZY Q-LEARNING ALGORITHM
    Glowaty, Grzegorz
    COMPUTER SCIENCE-AGH, 2005, 7 : 77 - 87
  • [5] A Comparative Study of Policies in Q-Learning for Foraging Tasks
    Mohan, Yogeswaran
    Ponnambalam, S. G.
    Inayat-Hussain, Jawaid I.
    2009 WORLD CONGRESS ON NATURE & BIOLOGICALLY INSPIRED COMPUTING (NABIC 2009), 2009, : 134 - +
  • [6] Convergence Improvement of Q-learning Based on a Personalized Recommendation System
    Chiang, Chia-Ling
    Cheng, Ming-Yang
    Ye, Ting-Yu
    Chen, Ya-Ling
    Huang, Pin-Hsuan
    2019 INTERNATIONAL AUTOMATIC CONTROL CONFERENCE (CACS), 2019,
  • [7] Optimizing Q-Learning with K-FAC AlgorithmOptimizing Q-Learning with K-FAC Algorithm
    Beltiukov, Roman
    ANALYSIS OF IMAGES, SOCIAL NETWORKS AND TEXTS (AIST 2019), 2020, 1086 : 3 - 8
  • [8] An Efficient Hardware Implementation of Reinforcement Learning: The Q-Learning Algorithm
    Spano, Sergio
    Cardarilli, Gian Carlo
    Di Nunzio, Luca
    Fazzolari, Rocco
    Giardino, Daniele
    Matta, Marco
    Nannarelli, Alberto
    Re, Marco
    IEEE ACCESS, 2019, 7 : 186340 - 186351
  • [9] Q-Learning in Dynamic Treatment Regimes With Misclassified Binary Outcome
    Liu, Dan
    He, Wenqing
    STATISTICS IN MEDICINE, 2024, 43 (30) : 5885 - 5897
  • [10] A Novel Ensemble Q-Learning Algorithm for Policy Optimization in Large-Scale Networks
    Bozkus, Talha
    Mitra, Urbashi
    FIFTY-SEVENTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, IEEECONF, 2023, : 1381 - 1386