Expert-Guided Deep Reinforcement Learning for Flexible Job Shop Scheduling Problem

被引:0
作者
Zhang, Wenqiang [1 ,2 ,3 ]
Geng, Huili [3 ]
Bao, Xuan [3 ]
Gen, Mitsuo [4 ]
Zhang, Guohui [5 ]
Deng, Miaolei [1 ,2 ]
机构
[1] Henan Univ Technol, Minist Educ, Key Lab Grain Informat Proc & Control, Zhengzhou 450001, Henan, Peoples R China
[2] Henan Univ Technol, Henan Key Lab Grain Photoelect Detect & Control, Zhengzhou 450001, Henan, Peoples R China
[3] Henan Univ Technol, Coll Informat Sci & Engn, Zhengzhou 450001, Henan, Peoples R China
[4] Tokyo Univ Sci, Fuzzy Log Syst Inst, Shinjuku City, Japan
[5] Zhengzhou Univ Aeronaut, Sch Management Engn, Zhengzhou 450046, Peoples R China
来源
BIO-INSPIRED COMPUTING: THEORIES AND APPLICATIONS, PT 2, BIC-TA 2023 | 2024年 / 2062卷
基金
中国国家自然科学基金;
关键词
Flexible job shop scheduling; Priority dispatching rules; Deep reinforcement learning; Graph neural network; Expert-guided; ALGORITHM;
D O I
10.1007/978-981-97-2275-4_4
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Flexible job shop scheduling (FJSP) is crucial for automated production, ensuring efficiency and flexibility. In recent years, deep reinforcement learning (DRL) has achieved success in solving sequence decision-making problems. However, the efficiency of the generated scheduling plans is often constrained by the dependence of most DRL algorithms on priority dispatching rules (PDR). In order to enable the agent trained in DRL to autonomously choose operations and machines, this paper proposes an expert-guided deep reinforcement learning framework (EGDRL). Based on the representation of scheduling states using a disjunctive graph and an operation-machine topology graph, a graph neural network (GNN) is used to capture the complex relationships between operations and machines. More importantly, in the early stages of training, this paper introduces expert-guided solutions by using PDR to guide the action selection of reinforcement learning, which greatly improves the quality of decision-making. Experimental results consistently show that the proposed method outperforms traditional PDRs and other DRL algorithm. Notably, this superiority is observed across various scales, including larger instances. Additionally, the method exhibits robust performance on instances not encountered during training.
引用
收藏
页码:50 / 60
页数:11
相关论文
共 17 条
[1]  
Brandimarte P., 1993, Annals of Operations Research, V41, P157, DOI 10.1007/BF02023073
[2]   JOB-SHOP SCHEDULING WITH MULTIPURPOSE MACHINES [J].
BRUCKER, P ;
SCHLIE, R .
COMPUTING, 1990, 45 (04) :369-375
[3]   A self-learning genetic algorithm based on reinforcement learning for flexible job-shop scheduling problem [J].
Chen, Ronghua ;
Yang, Bo ;
Li, Shi ;
Wang, Shilong .
COMPUTERS & INDUSTRIAL ENGINEERING, 2020, 149
[4]   Distributed policy search reinforcement learning for job-shop scheduling tasks [J].
Gabel, Thomas ;
Riedmiller, Martin .
INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2012, 50 (01) :41-61
[5]   A two-stage artificial bee colony algorithm scheduling flexible job-shop scheduling problem with new job insertion [J].
Gao, Kai Zhou ;
Suganthan, Ponnuthurai Nagaratnam ;
Chua, Tay Jin ;
Chong, Chin Soon ;
Cai, Tian Xiang ;
Pan, Qan Ke .
EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (21) :7652-7663
[6]  
HAUPT R, 1989, OR SPEKTRUM, V11, P3
[7]   MATHEMATICAL-PROGRAMMING FOR THE SCHEDULING PROBLEM WITH ALTERNATE PROCESS PLANS IN FMS [J].
JIANG, JC ;
HSIAO, WC .
COMPUTERS & INDUSTRIAL ENGINEERING, 1994, 27 (1-4) :15-18
[8]   Attention Models in Graphs: A Survey [J].
Lee, John Boaz ;
Rossi, Ryan A. ;
Kim, Sungchul ;
Ahmed, Nesreen K. ;
Koh, Eunyee .
ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2019, 13 (06)
[9]   Dynamic scheduling for flexible job shop with new job insertions by deep reinforcement learning [J].
Luo, Shu .
APPLIED SOFT COMPUTING, 2020, 91
[10]   ANALYSIS OF SCHEDULING RULES FOR AN FMS [J].
MONTAZERI, M ;
VANWASSENHOVE, LN .
INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 1990, 28 (04) :785-802