Fault-Tolerant Control of Programmable Logic Controller-Based Production Systems With Deep Reinforcement Learning

被引:7
作者
Zinn, Jonas [1 ]
Vogel-Heuser, Birgit [1 ]
Gruber, Marius [1 ]
机构
[1] Tech Univ Munich, Inst Automat & Informat Syst Mech Engn, D-85748 Garching, Germany
关键词
agent-based design; artificial intelligence; design automation; machine learning; robotic systems; RESTART STATES;
D O I
10.1115/1.4050624
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
Fault-tolerant control policies that automatically restart programable logic controller-based automated production system during fault recovery can increase system availability. This article provides a proof of concept that such policies can be synthesized with deep reinforcement learning. The authors specifically focus on systems with multiple end-effectors that are actuated in only one or two axes, commonly used for assembly and logistics tasks. Due to the large number of actuators in multi-end-effector systems and the limited possibilities to track workpieces in a single coordinate system, these systems are especially challenging to learn. This article demonstrates that a hierarchical multi-agent deep reinforcement learning approach together with a separate coordinate prediction module per agent can overcome these challenges. The evaluation of the suggested approach on the simulation of a small laboratory demonstrator shows that it is capable of restarting the system and completing open tasks as part of fault recovery.
引用
收藏
页数:11
相关论文
共 50 条
[31]   Adaptive PI Controller Based on a Reinforcement Learning Algorithm for Speed Control of a DC Motor [J].
Alejandro-Sanjines, Ulbio ;
Maisincho-Jivaja, Anthony ;
Asanza, Victor ;
Lorente-Leyva, Leandro L. ;
Peluffo-Ordonez, Diego H. .
BIOMIMETICS, 2023, 8 (05)
[32]   Deep Reinforcement Learning-Based Access Control for Buffer-Aided Relaying Systems With Energy Harvesting [J].
Zhang, Haodi ;
Zhan, Di ;
Zhang, Chen Jason ;
Wu, Kaishun ;
Liu, Ye ;
Luo, Sheng .
IEEE ACCESS, 2020, 8 (08) :145006-145017
[33]   On reliability of reinforcement learning based production scheduling systems: a comparative survey [J].
de Puiseau, Constantin Waubert ;
Meyes, Richard ;
Meisen, Tobias .
JOURNAL OF INTELLIGENT MANUFACTURING, 2022, 33 (04) :911-927
[34]   On reliability of reinforcement learning based production scheduling systems: a comparative survey [J].
Constantin Waubert de Puiseau ;
Richard Meyes ;
Tobias Meisen .
Journal of Intelligent Manufacturing, 2022, 33 :911-927
[35]   Algorithms for dynamic control of a deep-sea mining vehicle based on deep reinforcement learning [J].
Chen, Qihang ;
Yang, Jianmin ;
Zhao, Wenhua ;
Tao, Longbin ;
Mao, Jinghang ;
Li, Zhiyuan .
OCEAN ENGINEERING, 2024, 298
[36]   Performance of deep reinforcement learning algorithms in two-echelon inventory control systems [J].
Stranieri, Francesco ;
Stella, Fabio ;
Kouki, Chaaben .
INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2024, 62 (17) :6211-6226
[37]   Deep Deterministic Policy Gradient to Regulate Feedback Control Systems Using Reinforcement Learning [J].
Arshad, Jehangir ;
Khan, Ayesha ;
Aftab, Mariam ;
Hussain, Mujtaba ;
Rehman, Ateeq Ur ;
Ahmad, Shafiq ;
Al-Shayea, Adel M. ;
Shafiq, Muhammad .
CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 71 (01) :1153-1169
[38]   Dynamic Spectrum Sharing Based on Deep Reinforcement Learning in Mobile Communication Systems [J].
Liu, Sizhuang ;
Pan, Changyong ;
Zhang, Chao ;
Yang, Fang ;
Song, Jian .
SENSORS, 2023, 23 (05)
[39]   Modeling Deep Reinforcement Learning based Architectures for Cyber-Physical Systems [J].
Gatto, Nicola ;
Kusmenko, Evgeny ;
Rumpe, Bernhard .
2019 ACM/IEEE 22ND INTERNATIONAL CONFERENCE ON MODEL DRIVEN ENGINEERING LANGUAGES AND SYSTEMS COMPANION (MODELS-C 2019), 2019, :196-202
[40]   TCP-Drinc: Smart Congestion Control Based on Deep Reinforcement Learning [J].
Xiao, Kefan ;
Mao, Shiwen ;
Tugnait, Jitendra K. .
IEEE ACCESS, 2019, 7 :11892-11904