Personalized Treatment Policies with the Novel Buckley-James Q-Learning Algorithm

被引:3
|
作者
Lee, Jeongjin [1 ]
Kim, Jong-Min [2 ]
机构
[1] Ohio State Univ, Dept Stat, Columbus, OH 43210 USA
[2] Univ Minnesota Morris, Stat Discipline, Morris, MN 56267 USA
关键词
Q-learning; reinforcement learning; precision medicine; Buckley-James Method; survival analysis; DYNAMIC TREATMENT REGIMES; LINEAR-REGRESSION; SURVIVAL;
D O I
10.3390/axioms13040212
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This research paper presents the Buckley-James Q-learning (BJ-Q) algorithm, a cutting-edge method designed to optimize personalized treatment strategies, especially in the presence of right censoring. We critically assess the algorithm's effectiveness in improving patient outcomes and its resilience across various scenarios. Central to our approach is the innovative use of the survival time to impute the reward in Q-learning, employing the Buckley-James method for enhanced accuracy and reliability. Our findings highlight the significant potential of personalized treatment regimens and introduce the BJ-Q learning algorithm as a viable and promising approach. This work marks a substantial advancement in our comprehension of treatment dynamics and offers valuable insights for augmenting patient care in the ever-evolving clinical landscape.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Q-Learning Algorithm Based on Incremental RBF Network
    Hu Y.
    Li D.
    He Y.
    Han J.
    Jiqiren/Robot, 2019, 41 (05): : 562 - 573
  • [42] PENALIZED Q-LEARNING FOR DYNAMIC TREATMENT REGIMENS
    Song, Rui
    Wang, Weiwei
    Zeng, Donglin
    Kosorok, Michael R.
    STATISTICA SINICA, 2015, 25 (03) : 901 - 920
  • [43] Ramp Metering Control Based on the Q-Learning Algorithm
    Ivanjko, Edouard
    Necoska, Daniela Koltovska
    Greguric, Martin
    Vujic, Miroslav
    Jurkovic, Goran
    Mandzuka, Sadko
    CYBERNETICS AND INFORMATION TECHNOLOGIES, 2015, 15 (05) : 88 - 97
  • [44] Integrated Q-Learning with Firefly Algorithm for Transportation Problems
    Pratiba K.R.
    Ridhanya S.
    Ridhisha J.
    Hemashree P.
    EAI Endorsed Transactions on Energy Web, 2024, 11 : 1 - 6
  • [45] Sink Attraction Q-Learning Routing Algorithm For UWSNs
    Zhang, Zhi
    Li, Yibing
    Gao, Jialiang
    Ye, Fang
    2024 IEEE/CIC INTERNATIONAL CONFERENCE ON COMMUNICATIONS IN CHINA, ICCC, 2024,
  • [46] Discrete-Time Optimal Control Scheme Based on Q-Learning Algorithm
    Wei, Qinglai
    Liu, Derong
    Song, Ruizhuo
    2016 SEVENTH INTERNATIONAL CONFERENCE ON INTELLIGENT CONTROL AND INFORMATION PROCESSING (ICICIP), 2016, : 125 - 130
  • [47] Path planning for unmanned surface vehicle based on improved Q-Learning algorithm
    Wang, Yuanhui
    Lu, Changzhou
    Wu, Peng
    Zhang, Xiaoyue
    OCEAN ENGINEERING, 2024, 292
  • [48] An inverse reinforcement learning framework with the Q-learning mechanism for the metaheuristic algorithm
    Zhao, Fuqing
    Wang, Qiaoyun
    Wang, Ling
    KNOWLEDGE-BASED SYSTEMS, 2023, 265
  • [49] Robust Q-learning algorithm for Markov decision processes under Wasserstein uncertainty
    Neufeld, Ariel
    Sester, Julian
    AUTOMATICA, 2024, 168
  • [50] Adaptive Q-learning path planning algorithm based on virtual target guidance
    Li Z.
    Hu X.
    Zhang Y.
    Xu J.
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2024, 30 (02): : 553 - 568