Real-time measurement-driven reinforcement learning control approach for uncertain nonlinear systems

被引:8
作者
Abouheaf, Mohamed [1 ]
Boase, Derek [2 ]
Gueaieb, Wail [2 ]
Spinello, Davide [3 ]
Al-Sharhan, Salah [4 ]
机构
[1] Bowling Green State Univ, Coll Technol Architecture & Appl Engn, Bowling Green, OH 43403 USA
[2] Univ Ottawa, Sch Elect Engn & Comp Sci, 800 King Edward Ave, Ottawa, ON K1N 6N5, Canada
[3] Univ Ottawa, Dept Mech Engn, 161 Louis Pasteur, Ottawa, ON K1N 6N5, Canada
[4] Int Univ Sci & Technol, Comp Engn Dept, Ardiya, Kuwait
基金
加拿大自然科学与工程研究理事会;
关键词
Optimal control; Adaptive control; Reinforcement learning; Adaptive critics; Model-reference adaptive systems;
D O I
10.1016/j.engappai.2023.106029
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The paper introduces an interactive machine learning mechanism to process the measurements of an uncertain, nonlinear dynamic process and hence advise an actuation strategy in real-time. For concept demonstration, a trajectory-following optimization problem of a Kinova robotic arm is solved using an integral reinforcement learning approach with guaranteed stability for slowly varying dynamics. The solution is implemented using a model-free value iteration process to solve the integral temporal difference equations of the problem. The performance of the proposed technique is benchmarked against that of another model-free high-order approach and is validated for dynamic payload and disturbances. Unlike its benchmark, the proposed adaptive strategy is capable of handling extreme process variations. This is experimentally demonstrated by introducing static and time-varying payloads close to the rated maximum payload capacity of the manipulator arm. The comparison algorithm exhibited up to a seven-fold percent overshoot compared to the proposed integral reinforcement learning solution. The robustness of the algorithm is further validated by disturbing the real-time adapted strategy gains with a white noise of a standard deviation as high as 5%.
引用
收藏
页数:18
相关论文
共 58 条
  • [1] Discrete-time dynamic graphical games: model-free reinforcement learning solution
    Abouheaf M.I.
    Lewis F.L.
    Mahmoud M.S.
    Mikulski D.G.
    [J]. Control theory technol., 1 (55-69): : 55 - 69
  • [2] A Data-Driven Model-Reference Adaptive Control Approach Based on Reinforcement Learning
    Abouheaf, Mohammed
    Gueaieb, Wail
    Spinello, Davide
    Al-Sharhan, Salah
    [J]. 2021 IEEE INTERNATIONAL SYMPOSIUM ON ROBOTIC AND SENSORS ENVIRONMENTS (ROSE 2021), 2021,
  • [3] Guidance Mechanism for Flexible-Wing Aircraft Using Measurement-Interfaced Machine-Learning Platform
    Abouheaf, Mohammed
    Mailhot, Nathaniel Q.
    Gueaieb, Wail
    Spinello, Davide
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2020, 69 (07) : 4637 - 4648
  • [4] Abouheaf M, 2019, 2019 IEEE INTERNATIONAL SYMPOSIUM ON ROBOTIC AND SENSORS ENVIRONMENTS (ROSE 2019), P84, DOI [10.1109/rose.2019.8790432, 10.1109/rose.2019.8790424]
  • [5] Load frequency regulation for multi-area power system using integral reinforcement learning
    Abouheaf, Mohammed
    Gueaieb, Wail
    Sharaf, Adel
    [J]. IET GENERATION TRANSMISSION & DISTRIBUTION, 2019, 13 (19) : 4311 - 4323
  • [6] Abouheaf M, 2019, IEEE INT CONF ROBOT, P2195, DOI [10.1109/ICRA.2019.8794438, 10.1109/icra.2019.8794438]
  • [7] Multi-agent discrete-time graphical games and reinforcement learning solutions
    Abouheaf, Mohammed I.
    Lewis, Frank L.
    Vamvoudakis, Kyriakos G.
    Haesaert, Sofie
    Babuska, Robert
    [J]. AUTOMATICA, 2014, 50 (12) : 3038 - 3053
  • [8] Autonomous Unmanned Aerial Vehicle navigation using Reinforcement Learning: A systematic review
    AlMahamid, Fadi
    Grolinger, Katarina
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 115
  • [9] Sleep Apnea Detection From Single-Lead ECG: A Comprehensive Analysis of Machine Learning and Deep Learning Algorithms
    Bahrami, Mahsa
    Forouzanfar, Mohamad
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71
  • [10] Ben Amor R, 2017, I C SCI TECH AUTO CO, P235, DOI 10.1109/STA.2017.8314907