Digital twin-based reinforcement learning framework: application to autonomous mobile robot dispatching

被引:5
作者
Jaoua, Amel [1 ]
Masmoudi, Samar [1 ]
Negri, Elisa [2 ]
机构
[1] Univ Tunis El Manar, Natl Engn Sch Tunis, Ind Engn, LR OASIS, Tunis, Tunisia
[2] Politecn Milan, Management Econ & Ind Engn, Milan, Italy
关键词
Digital twin; reinforcement learning; deep Q-Network; real-time; dispatching; autonomous mobile robot; SIMULATION; OPTIMIZATION;
D O I
10.1080/0951192X.2024.2314787
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
This paper proposes a new framework for embedding an Intelligent Digital Twin (DT) in a production system with the objective of achieving more efficient real-time production planning and control. For that purpose, the Intelligence Layer is based on Reinforcement Leaning (RL) and Deep RL (DRL) algorithms. The use of this control instead of parametric simulation-based optimization approach allows to benefit from the separation between the training and execution phase. To ensure consistency and reusability, this work presents a standardized framework, based on a formal methodology, that specifies how the various components of the DT-based RL architecture interact over time to achieve essential real-time concurrency and synchronization aspects. Experiments are conducted in a small-scale production system where material handling operations are performed by an Autonomous Mobile Robot (AMR) in an Industry 4.0 Laboratory. Results showed how synchronized state updates between the Physical and Cyber World are used within the Decision Layer to ensure real-time response for the AMR dispatching requests. Finally, to deal with continuous and high-dimensional state spaces, the Deep Q-Network is implemented. The findings of an extensive computational study reveal that implementing the DT-based DRL solution leads to improved efficiency and robustness when compared to conventional dispatching rules.
引用
收藏
页码:1335 / 1358
页数:24
相关论文
共 48 条
[41]   (Data-Driven) Development of dynamic scheduling in semiconductor manufacturing using a Q-learning approach [J].
Shiue, Yeou-Ren ;
Lee, Ken-Chuan ;
Su, Chao-Ton .
INTERNATIONAL JOURNAL OF COMPUTER INTEGRATED MANUFACTURING, 2022, 35 (10-11) :1188-1204
[42]   Digital Twin: Origin to Future [J].
Singh, Maulshree ;
Fuenmayor, Evert ;
Hinchy, Eoin P. ;
Qiao, Yuansong ;
Murray, Niall ;
Devine, Declan .
APPLIED SYSTEM INNOVATION, 2021, 4 (02)
[43]  
Sutton RS, 2018, ADAPT COMPUT MACH LE, P1
[44]   Next generation DES simulation: A research agenda for human centric manufacturing systems [J].
Turner, Chris J. ;
Garn, Wolfgang .
JOURNAL OF INDUSTRIAL INFORMATION INTEGRATION, 2022, 28
[45]   Perspective on holonic manufacturing systems: PROSA becomes ARTI [J].
Valckenaers, Paul .
COMPUTERS IN INDUSTRY, 2020, 120
[46]   A decision-making framework for dynamic scheduling of cyber-physical production systems based on digital twins [J].
Villalonga, Alberto ;
Negri, Elisa ;
Biscardo, Giacomo ;
Castano, Fernando ;
Haber, Rodolfo E. ;
Fumagalli, Luca ;
Macchi, Marco .
ANNUAL REVIEWS IN CONTROL, 2021, 51 :357-373
[47]   Adaptive scheduling for assembly job shop with uncertain assembly times based on dual Q-learning [J].
Wang, Haoxiang ;
Sarker, Bhaba R. ;
Li, Jing ;
Li, Jian .
INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2021, 59 (19) :5867-5883
[48]   Reinforcement learning and digital twin-based real-time scheduling method in intelligent manufacturing systems [J].
Zhang, Lixiang ;
Yan, Yan ;
Hu, Yaoguang ;
Ren, Weibo .
IFAC PAPERSONLINE, 2022, 55 (10) :359-364