Hybrid Control Policy for Artificial Pancreas via Ensemble Deep Reinforcement Learning

被引:0
|
作者
Lv, Wenzhou [1 ]
Wu, Tianyu [1 ]
Xiong, Luolin [1 ]
Wu, Liang [2 ,3 ]
Zhou, Jian [2 ,3 ]
Tang, Yang [1 ]
Qian, Feng [1 ]
机构
[1] East China Univ Sci & Technol, State Key Lab Ind Control Technol, Shanghai 200237, Peoples R China
[2] Shanghai Jiao Tong Univ, Metab, Shanghai, Peoples R China
[3] Shanghai Diabet Inst, Shanghai Clin Ctr Diabet, Peoples Hosp 6, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Glucose; Insulin; Safety; Uncertainty; Pancreas; Metalearning; Accuracy; Artificial pancreas; glucose control; diabetes; reinforcement learning; meta learning; TYPE-1; MPC; SAFETY;
D O I
10.1109/TBME.2024.3451712
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Objective: The artificial pancreas (AP) shows promise for closed-loop glucose control in type 1 diabetes mellitus (T1DM). However, designing effective control policies for the AP remains challenging due to complex physiological processes, delayed insulin response, and inaccurate glucose measurements. While model predictive control (MPC) offers safety and stability through the dynamic model and safety constraints, it lacks individualization and is adversely affected by unannounced meals. Conversely, deep reinforcement learning (DRL) provides personalized and adaptive strategies but struggles with distribution shifts and substantial data requirements. Methods: We propose a hybrid control policy for the artificial pancreas (HyCPAP) to address the above challenges. HyCPAP combines an MPC policy with an ensemble DRL policy, leveraging the strengths of both policies while compensating for their respective limitations. To facilitate faster deployment of AP systems in real-world settings, we further incorporate meta-learning techniques into HyCPAP, leveraging previous experience and patient-shared knowledge to enable fast adaptation to new patients with limited available data. Results: We conduct extensive experiments using the UVA/Padova T1DM simulator across five scenarios. Our approaches achieve the highest percentage of time spent in the desired range and the lowest occurrences of hypoglycemia. Conclusion: The results clearly demonstrate the superiority of our methods for closed-loop glucose management in individuals with T1DM. Significance: The study presents novel control policies for AP systems, affirming their great potential for efficient closed-loop glucose control.
引用
收藏
页码:309 / 323
页数:15
相关论文
共 50 条
  • [41] Learning to Walk via Deep Reinforcement Learning
    Haarnoja, Tuomas
    Ha, Sehoon
    Zhou, Aurick
    Tan, Jie
    Tucker, George
    Levine, Sergey
    ROBOTICS: SCIENCE AND SYSTEMS XV, 2019,
  • [42] Bayesian Deep Reinforcement Learning via Deep Kernel Learning
    Junyu Xuan
    Jie Lu
    Zheng Yan
    Guangquan Zhang
    International Journal of Computational Intelligence Systems, 2018, 12 : 164 - 171
  • [43] Bayesian Deep Reinforcement Learning via Deep Kernel Learning
    Xuan, Junyu
    Lu, Jie
    Yan, Zheng
    Zhang, Guangquan
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2019, 12 (01) : 164 - 171
  • [44] Intelligent H∞ Control for UAVs via Fuzzy Deep Reinforcement Learning
    Cheng, Haoyu
    Wang, Meng
    Ma, Yifeng
    Jiao, Jiayue
    Song, Ruijia
    PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, : 7182 - 7187
  • [45] Hybrid UAV-Enabled Secure Offloading via Deep Reinforcement Learning
    Yoo, Seonghoon
    Jeong, Seongah
    Kang, Joonhyuk
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2023, 12 (06) : 972 - 976
  • [46] District-Coupled Epidemic Control via Deep Reinforcement Learning
    Du, Xinqi
    Liu, Tianyi
    Zhao, Songwei
    Song, Jiuman
    Chen, Hechang
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT II, 2022, 13369 : 417 - 428
  • [47] Intelligent Control for Unmanned Flight Vehicles via Deep Reinforcement Learning
    Cheng, Haoyu
    Zhang, Xiaofeng
    Huang, Hanqiao
    Zhao, Xiaohan
    PROCEEDINGS OF THE 32ND 2020 CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2020), 2020, : 3184 - 3189
  • [48] Dynamic Channel Access and Power Control via Deep Reinforcement Learning
    Lu, Ziyang
    Gursoy, M. Cenk
    2019 IEEE 90TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2019-FALL), 2019,
  • [49] Variable Admittance Interaction Control of UAVs via Deep Reinforcement Learning
    Feng, Yuting
    Shi, Chuanbeibei
    Du, Jianrui
    Yu, Yushu
    Sun, Fuchun
    Song, Yixu
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 1291 - 1297
  • [50] Hierarchical Active Tracking Control for UAVs via Deep Reinforcement Learning
    Zhao, Wenlong
    Meng, Zhijun
    Wang, Kaipeng
    Zhang, Jiahui
    Lu, Shaoze
    APPLIED SCIENCES-BASEL, 2021, 11 (22):