Deep Reinforcement Learning-Based Job Shop Scheduling of Smart Manufacturing

被引:14
作者
Elsayed, Eman K. [1 ]
Elsayed, Asmaa K. [2 ]
Eldahshan, Kamal A. [3 ]
机构
[1] AL Azhar Univ, Fac Sci Girls, Dept Math, Sch Comp Sci,Canadian Int Coll CIC, Cairo 11511, Egypt
[2] AL Azhar Univ, Fac Sci Girls, Dept Math, Cairo 11511, Egypt
[3] AL Azhar Univ, Fac Sci, Dept Math, Cairo 11511, Egypt
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2022年 / 73卷 / 03期
关键词
Reinforcement learning; job shop scheduling; graphical isomorphism network; actor-critic networks; MACHINE;
D O I
10.32604/cmc.2022.030803
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Industry 4.0 production environments and smart manufacturing systems integrate both the physical and decision-making aspects of manufacturing operations into autonomous and decentralized systems. One of the key aspects of these systems is a production planning, specifically, Scheduling operations on the machines. To cope with this problem, this paper proposed a Deep Reinforcement Learning with an Actor-Critic algorithm (DRLAC). We model the Job-Shop Scheduling Problem (JSSP) as a Markov Decision Process (MDP), represent the state of a JSSP as simple Graph Isomorphism Networks (GIN) to extract nodes features during scheduling, and derive the policy of optimal scheduling which guides the included node features to the best next action of schedule. In addition, we adopt the Actor-Critic (AC) network's training algorithm-based reinforcement learning for achieving the optimal policy of the scheduling. To prove the proposed model's effectiveness, first, we will present a case study that illustrated a conflict between two job scheduling, secondly, we will apply the proposed model to a known benchmark dataset and compare the results with the traditional scheduling methods and trending approaches. The numerical results indicate that the proposed model can be adaptive with real-time production scheduling, where the average percentage deviation (APD) of our model achieved values between 0.009 and 0.21 compared with heuristic methods and values between 0.014 and 0.18 compared with other trending approaches.
引用
收藏
页码:5103 / 5120
页数:18
相关论文
共 39 条
[1]  
Abdolrazzagh-Nezhad M., 2017, INT J COMPUT INF ENG, V11, P429
[2]   Deep Reinforcement Learning Model for Blood Bank Vehicle Routing Multi-Objective Optimization [J].
Altaf, Meteb M. ;
Roshdy, Ahmed Samir ;
AlSagri, Hatoon S. .
CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 70 (02) :3955-3967
[3]  
[Anonymous], JOBSHOP INSTANCES
[4]  
[Anonymous], Github Repositories
[5]  
Applegate D., 1991, ORSA Journal on Computing, V3, P149, DOI 10.1287/ijoc.3.2.149
[6]   A multistage graph-based procedure for solving a just-in-time flexible job-shop scheduling problem with machine and time-dependent processing costs [J].
Corominas, Albert ;
Garcia-Villoria, Alberto ;
Gonzalez, Nestor-Andres ;
Pastor, Rafael .
JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 2019, 70 (04) :620-633
[7]  
Dasoulas G., 2020, COLORING GRAPH NEURA, P2098
[8]   Minimizing the Late Work of the Flow Shop Scheduling Problem with a Deep Reinforcement Learning Based Approach [J].
Dong, Zhuoran ;
Ren, Tao ;
Weng, Jiacheng ;
Qi, Fang ;
Wang, Xinyue .
APPLIED SCIENCES-BASEL, 2022, 12 (05)
[9]  
Elsayed Asmaa K., 2021, 2021 Tenth International Conference on Intelligent Computing and Information Systems (ICICIS), P32, DOI 10.1109/ICICIS52592.2021.9694207
[10]  
Errica F., 2022, ARXIV PREPRINT ARXIV, V3, P1