Reliability evaluation of reinforcement learning methods for mechanical systems with increasing complexity

被引:4
|
作者
Manzl, Peter [1 ]
Rogov, Oleg [2 ]
Gerstmayr, Johannes [1 ]
Mikkola, Aki [2 ]
Orzechowski, Grzegorz [2 ]
机构
[1] Univ Innsbruck, Dept Mechatron, Tech Str 13, A-6020 Innsbruck, Tyrol, Austria
[2] LUT Univ, Dept Mech Engn, Yliopistonkatu 34, Lappeenranta 53850, South Karelia, Finland
关键词
Reinforcement learning; Reliability analysis; Inverse pendulum; Machine learning; Dynamical systems; DYNAMICS;
D O I
10.1007/s11044-023-09960-2
中图分类号
O3 [力学];
学科分类号
08 ; 0801 ;
摘要
Reinforcement learning (RL) is one of the emerging fields of artificial intelligence (AI) intended for designing agents that take actions in the physical environment. RL has many vital applications, including robotics and autonomous vehicles. The key characteristic of RL is its ability to learn from experience without requiring direct programming or supervision. To learn, an agent interacts with an environment by acting and observing the resulting states and rewards. In most practical applications, an environment is implemented as a virtual system due to cost, time, and safety concerns. Simultaneously, multibody system dynamics (MSD) is a framework for efficiently and systematically developing virtual systems of arbitrary complexity. MSD is commonly used to create virtual models of robots, vehicles, machinery, and humans. The features of RL and MSD make them perfect companions in building sophisticated, automated, and autonomous mechatronic systems. The research demonstrates the use of RL in controlling multibody systems. While AI methods are used to solve some of the most challenging tasks in engineering, their proper understanding and implementation are demanding. Therefore, we introduce and detail three commonly used RL algorithms to control the inverted N-pendulum on the cart. Single-, double-, and triple-pendulum configurations are investigated, showing the capability of RL methods to handle increasingly complex dynamical systems. We show 2D state space zones where the agent succeeds or fails the stabilization. Despite passing randomized tests during training, blind spots may occur where the agent's policy fails. Results confirm that RL is a versatile, although complex, control engineering approach.
引用
收藏
页数:25
相关论文
共 50 条
  • [21] Reinforcement learning in reliability and maintenance optimization: A tutorial
    Zhang, Qin
    Liu, Yu
    Xiang, Yisha
    Xiahou, Tangfan
    RELIABILITY ENGINEERING & SYSTEM SAFETY, 2024, 251
  • [22] A Low-Complexity Algorithm for a Reinforcement Learning-Based Channel Estimator for MIMO Systems
    Kim, Tae-Kyoung
    Min, Moonsik
    SENSORS, 2022, 22 (12)
  • [23] Reinforcement Learning in BitTorrent Systems
    Izhak-Ratzin, Rafit
    Park, Hyunggon
    van der Schaar, Mihaela
    2011 PROCEEDINGS IEEE INFOCOM, 2011, : 406 - 410
  • [24] Large-scale prediction of activity cliffs using machine and deep learning methods of increasing complexity
    Shunsuke Tamura
    Tomoyuki Miyao
    Jürgen Bajorath
    Journal of Cheminformatics, 15
  • [25] Large-scale prediction of activity cliffs using machine and deep learning methods of increasing complexity
    Tamura, Shunsuke
    Miyao, Tomoyuki
    Bajorath, Juergen
    JOURNAL OF CHEMINFORMATICS, 2023, 15 (01)
  • [26] Using Localization and Factorization to Reduce the Complexity of Reinforcement Learning
    Sunehag, Peter
    Hutter, Marcus
    ARTIFICIAL GENERAL INTELLIGENCE (AGI 2015), 2015, 9205 : 177 - 186
  • [27] Low Complexity SON Coordination using Reinforcement Learning
    Iacoboaiea, Ovidiu
    Sayrac, Berna
    Ben Jemaa, Sana
    Bianchi, Pascal
    2014 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2014), 2014, : 4406 - 4411
  • [28] Complexity Analysis of Reinforcement Learning and Its Application to Robotics
    Li, Bocheng
    Xia, Li
    Zhao, Qianchuan
    2017 13TH IEEE CONFERENCE ON AUTOMATION SCIENCE AND ENGINEERING (CASE), 2017, : 1425 - 1426
  • [29] Speedy Categorical Distributional Reinforcement Learning and Complexity Analysis
    Boeck, Markus
    Heitzinger, Clemens
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2022, 4 (02): : 675 - 693
  • [30] Symbolic Regression Methods for Reinforcement Learning
    Kubalik, Jiri
    Derner, Erik
    Zegklitz, Jan
    Babuska, Robert
    IEEE ACCESS, 2021, 9 : 139697 - 139711