Deep Reinforcement Learning for Guidewire Navigation in Coronary Artery Phantom

被引:20
|
作者
Kweon, Jihoon [1 ]
Kim, Kyunghwan [2 ]
Lee, Chaehyuk [2 ]
Kwon, Hwi [2 ]
Park, Jinwoo [2 ]
Song, Kyoseok [2 ]
Kim, Young In [3 ]
Park, Jeeone [3 ]
Back, Inwook [4 ]
Roh, Jae-Hyung [5 ]
Moon, Youngjin [1 ]
Choi, Jaesoon [6 ]
Kim, Young-Hak [4 ]
机构
[1] Univ Ulsan, Coll Med, Asan Med Ctr, Dept Convergence Med, Seoul 05505, South Korea
[2] Medipixel Inc, Seoul 04037, South Korea
[3] Univ Ulsan, Coll Med, Asan Med Inst Convergence Sci & Technol, Dept Med Sci,Asan Med Ctr, Seoul 05505, South Korea
[4] Univ Ulsan, Coll Med, Asan Med Ctr, Div Cardiol,Dept Internal Med, Seoul 05505, South Korea
[5] Chungnam Natl Univ, Sch Med, Sejong Hosp, Dept Cardiol Internal Med, Daejeon 30099, South Korea
[6] Univ Ulsan, Coll Med, Asan Med Ctr, Dept Biomed Engn, Seoul 05505, South Korea
关键词
Training; Robots; Navigation; Phantoms; Arteries; Lesions; Reinforcement learning; Coronary intervention; guidewire navigation; reinforcement learning; INTERVENTION; FEASIBILITY; SAFETY;
D O I
10.1109/ACCESS.2021.3135277
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In percutaneous intervention for treatment of coronary plaques, guidewire navigation is a primary procedure for stent delivery. Steering a flexible guidewire within coronary arteries requires considerable training, and the non-linearity between the control operation and the movement of the guidewire makes precise manipulation difficult. Here, we introduce a deep reinforcement learning (RL) framework for autonomous guidewire navigation in a robot-assisted coronary intervention. Using Rainbow, a segment-wise learning approach is applied to determine how best to accelerate training using human demonstrations, transfer learning, and weight initialization. 'State' for RL is customized as a focus window near the guidewire tip, and subgoals are placed to mitigate a sparse reward problem. The RL agent improves performance, eventually enabling the guidewire to reach all valid targets in 'stable' phase. For the last 300 out of 1000 episodes, the success rates of the guidewire navigation to the distal-main and side targets were 98% and 99% in 2D and 3D phantoms, respectively. Our framework opens a new direction in the automation of robot-assisted intervention, providing guidance on RL in physical spaces involving mechanical fatigue.
引用
收藏
页码:166409 / 166422
页数:14
相关论文
共 50 条
  • [1] Cooperative Deep Reinforcement Learning Policies for Autonomous Navigation in Complex Environments
    Tran, Van Manh
    Kim, Gon-Woo
    IEEE ACCESS, 2024, 12 : 101053 - 101065
  • [2] Tutor-Guided Interior Navigation With Deep Reinforcement Learning
    Zeng, Fanyu
    Wang, Chen
    Ge, Shuzhi Sam
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2021, 13 (04) : 934 - 944
  • [3] Visual Navigation With Multiple Goals Based on Deep Reinforcement Learning
    Rao, Zhenhuan
    Wu, Yuechen
    Yang, Zifei
    Zhang, Wei
    Lu, Shijian
    Lu, Weizhi
    Zha, ZhengJun
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (12) : 5445 - 5455
  • [4] Multigoal Visual Navigation With Collision Avoidance via Deep Reinforcement Learning
    Xiao, Wendong
    Yuan, Liang
    He, Li
    Ran, Teng
    Zhang, Jianbo
    Cui, Jianping
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71
  • [5] A zero-shot reinforcement learning strategy for autonomous guidewire navigation
    Scarponi, Valentina
    Duprez, Michel
    Nageotte, Florent
    Cotin, Stephane
    INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2024, 19 (06) : 1185 - 1192
  • [6] Using Deep Reinforcement Learning for Navigation in Simulated Hallways
    Leao, Goncalo
    Almeida, Filipe
    Trigo, Emanuel
    Ferreira, Henrique
    Sousa, Armando
    Reis, Luis Paulo
    2023 IEEE INTERNATIONAL CONFERENCE ON AUTONOMOUS ROBOT SYSTEMS AND COMPETITIONS, ICARSC, 2023, : 207 - 213
  • [7] Indoor Navigation with Deep Reinforcement Learning
    Bakale, Vijayalakshmi A.
    Kumar, Yeshwanth V. S.
    Roodagi, Vivekanand C.
    Kulkarni, Yashaswini N.
    Patil, Mahesh S.
    Chickerur, Satyadhyan
    PROCEEDINGS OF THE 5TH INTERNATIONAL CONFERENCE ON INVENTIVE COMPUTATION TECHNOLOGIES (ICICT-2020), 2020, : 660 - 665
  • [8] Hindsight Intermediate Targets for Mapless Navigation With Deep Reinforcement Learning
    Jang, Yoonsu
    Baek, Jongchan
    Han, Soohee
    AMERICAN JOURNAL OF CLINICAL NUTRITION, 2021, : 11816 - 11825
  • [9] Path-Following Navigation in Crowds With Deep Reinforcement Learning
    Fu, Hao
    Wang, Qiang
    He, Haodong
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (11): : 20236 - 20245
  • [10] Hindsight Intermediate Targets for Mapless Navigation With Deep Reinforcement Learning
    Jang, Yoonsu
    Baek, Jongchan
    Han, Soohee
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2022, 69 (11) : 11816 - 11825