Personalized Treatment Policies with the Novel Buckley-James Q-Learning Algorithm

被引:3
|
作者
Lee, Jeongjin [1 ]
Kim, Jong-Min [2 ]
机构
[1] Ohio State Univ, Dept Stat, Columbus, OH 43210 USA
[2] Univ Minnesota Morris, Stat Discipline, Morris, MN 56267 USA
关键词
Q-learning; reinforcement learning; precision medicine; Buckley-James Method; survival analysis; DYNAMIC TREATMENT REGIMES; LINEAR-REGRESSION; SURVIVAL;
D O I
10.3390/axioms13040212
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This research paper presents the Buckley-James Q-learning (BJ-Q) algorithm, a cutting-edge method designed to optimize personalized treatment strategies, especially in the presence of right censoring. We critically assess the algorithm's effectiveness in improving patient outcomes and its resilience across various scenarios. Central to our approach is the innovative use of the survival time to impute the reward in Q-learning, employing the Buckley-James method for enhanced accuracy and reliability. Our findings highlight the significant potential of personalized treatment regimens and introduce the BJ-Q learning algorithm as a viable and promising approach. This work marks a substantial advancement in our comprehension of treatment dynamics and offers valuable insights for augmenting patient care in the ever-evolving clinical landscape.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] A Novel Q-learning Algorithm with Function Approximation for Constrained Markov Decision Processes
    Lakshmanan, K.
    Bhatnagar, Shalabh
    2012 50TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2012, : 400 - 405
  • [22] Generating Test Cases for Q-Learning Algorithm
    Kumaresan, Lavanya
    Chamundeswari, A.
    2013 FOURTH INTERNATIONAL CONFERENCE ON COMPUTING, COMMUNICATIONS AND NETWORKING TECHNOLOGIES (ICCCNT), 2013,
  • [23] A novel Q-Learning Algorithm Based on the Stochastic Environment Path Planning Problem
    Jian, Li
    Rong, Fei
    Yu, Tang
    2020 IEEE 19TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2020), 2020, : 1977 - 1982
  • [24] Extended Q-Learning Algorithm for Path-Planning of a Mobile Robot
    Goswami , Indrani
    Das, Pradipta Kumar
    Konar, Amit
    Janarthanan, R.
    SIMULATED EVOLUTION AND LEARNING, 2010, 6457 : 379 - +
  • [25] Guidance law based on zero effort miss and Q-learning algorithm
    He, Xianjun
    Chen, Zhihua
    Jia, Fang
    Wu, Mingyu
    SEVENTH SYMPOSIUM ON NOVEL PHOTOELECTRONIC DETECTION TECHNOLOGY AND APPLICATIONS, 2021, 11763
  • [26] Study of Cooperation Strategy of Robot Based on Parallel Q-Learning Algorithm
    Wang, Shuda
    Si, Feng
    Yang, Jing
    Wang, Shuoning
    Yang, Jun
    INTELLIGENT ROBOTICS AND APPLICATIONS, PT I, PROCEEDINGS, 2008, 5314 : 633 - 642
  • [27] An improved Q-learning algorithm for an autonomous mobile robot navigation problem
    Muhammad, Jawad
    Bucak, Ihsan Omur
    2013 INTERNATIONAL CONFERENCE ON TECHNOLOGICAL ADVANCES IN ELECTRICAL, ELECTRONICS AND COMPUTER ENGINEERING (TAEECE), 2013, : 239 - 243
  • [28] Solving a Job Shop Scheduling Problem Using Q-Learning Algorithm
    Belmamoune, Manal Abir
    Ghomri, Latefa
    Yahouni, Zakaria
    12TH INTERNATIONAL WORKSHOP ON SERVICE ORIENTED, HOLONIC AND MULTI-AGENT MANUFACTURING SYSTEMS FOR INDUSTRY OF THE FUTURE, SOHOMA 2022, 2023, 1083 : 196 - 209
  • [29] Efficient Q-learning hyperparameter tuning using FOX optimization algorithm
    Jumaah, Mahmood A.
    Ali, Yossra H.
    Rashid, Tarik A.
    RESULTS IN ENGINEERING, 2025, 25
  • [30] Elevator group control algorithm based on residual gradient and Q-learning
    Zong, ZL
    Wang, XG
    Tang, Z
    Zeng, GZ
    SICE 2004 ANNUAL CONFERENCE, VOLS 1-3, 2004, : 329 - 331