Personalized Treatment Policies with the Novel Buckley-James Q-Learning Algorithm

被引:3
|
作者
Lee, Jeongjin [1 ]
Kim, Jong-Min [2 ]
机构
[1] Ohio State Univ, Dept Stat, Columbus, OH 43210 USA
[2] Univ Minnesota Morris, Stat Discipline, Morris, MN 56267 USA
关键词
Q-learning; reinforcement learning; precision medicine; Buckley-James Method; survival analysis; DYNAMIC TREATMENT REGIMES; LINEAR-REGRESSION; SURVIVAL;
D O I
10.3390/axioms13040212
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This research paper presents the Buckley-James Q-learning (BJ-Q) algorithm, a cutting-edge method designed to optimize personalized treatment strategies, especially in the presence of right censoring. We critically assess the algorithm's effectiveness in improving patient outcomes and its resilience across various scenarios. Central to our approach is the innovative use of the survival time to impute the reward in Q-learning, employing the Buckley-James method for enhanced accuracy and reliability. Our findings highlight the significant potential of personalized treatment regimens and introduce the BJ-Q learning algorithm as a viable and promising approach. This work marks a substantial advancement in our comprehension of treatment dynamics and offers valuable insights for augmenting patient care in the ever-evolving clinical landscape.
引用
收藏
页数:12
相关论文
共 50 条
  • [11] A new Q-learning algorithm based on the Metropolis criterion
    Guo, MZ
    Liu, Y
    Malec, J
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2004, 34 (05): : 2140 - 2143
  • [12] A Weighted Smooth Q-Learning Algorithm
    Vijesh, V. Antony
    Shreyas, S. R.
    IEEE CONTROL SYSTEMS LETTERS, 2025, 9 : 21 - 26
  • [13] ETQ-learning: an improved Q-learning algorithm for path planning
    Wang, Huanwei
    Jing, Jing
    Wang, Qianlv
    He, Hongqi
    Qi, Xuyan
    Lou, Rui
    INTELLIGENT SERVICE ROBOTICS, 2024, 17 (04) : 915 - 929
  • [14] A selection hyper-heuristic algorithm with Q-learning mechanism
    Zhao, Fuqing
    Liu, Yuebao
    Zhu, Ningning
    Xu, Tianpeng
    Jonrinaldi
    APPLIED SOFT COMPUTING, 2023, 147
  • [15] Heuristically accelerated Q-learning algorithm based on Laplacian Eigenmap
    Zhu, Mei-Qiang
    Li, Ming
    Cheng, Yu-Hu
    Zhang, Qian
    Wang, Xue-Song
    Kongzhi yu Juece/Control and Decision, 2014, 29 (03): : 425 - 430
  • [16] A study on a Q-Learning algorithm application to a manufacturing assembly problem
    Neves, Miguel
    Vieira, Miguel
    Neto, Pedro
    JOURNAL OF MANUFACTURING SYSTEMS, 2021, 59 : 426 - 440
  • [17] A New Algorithm to Track Dynamic Goal Position in Q-learning
    Mitra, Soumishila
    Banerjee, Dhrubojyoti
    Konar, Amit
    Janarthanan, R.
    2012 12TH INTERNATIONAL CONFERENCE ON HYBRID INTELLIGENT SYSTEMS (HIS), 2012, : 69 - 74
  • [18] A Path Planning Algorithm for UAV Based on Improved Q-Learning
    Yan, Chao
    Xiang, Xiaojia
    2018 2ND INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION SCIENCES (ICRAS), 2018, : 46 - 50
  • [19] A Path Planning Algorithm for Space Manipulator Based on Q-Learning
    Li, Taiguo
    Li, Quanhong
    Li, Wenxi
    Xia, Jiagao
    Tang, Wenhua
    Wang, Weiwen
    PROCEEDINGS OF 2019 IEEE 8TH JOINT INTERNATIONAL INFORMATION TECHNOLOGY AND ARTIFICIAL INTELLIGENCE CONFERENCE (ITAIC 2019), 2019, : 1566 - 1571
  • [20] A robot demonstration method based on LWR and Q-learning algorithm
    Zhao, Guangzhe
    Tao, Yong
    Liu, Hui
    Deng, Xianling
    Chen, Youdong
    Xiong, Hegen
    Xie, Xianwu
    Fang, Zengliang
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2018, 35 (01) : 35 - 46