Deep reinforcement learning driven trajectory-based meta-heuristic for distributed heterogeneous flexible job shop scheduling problem

被引:1
|
作者
Zhang, Qichen [1 ]
Shao, Weishi [1 ,3 ,4 ]
Shao, Zhongshi [2 ]
Pi, Dechang [4 ]
Gao, Jiaquan [1 ,3 ]
机构
[1] Nanjing Normal Univ, Sch Comp & Elect Informat, Sch Artificial Intelligence, Nanjing, Peoples R China
[2] Shaanxi Normal Univ, Sch Comp Sci, Xian, Peoples R China
[3] Minist Educ, Key Lab Numer Simulat Large Scale Complex Syst, Beijing, Peoples R China
[4] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing, Peoples R China
基金
中国博士后科学基金;
关键词
Distributed heterogeneous flexible job shop; scheduling problem; Deep Q network; Variable neighborhood search; Makespan; Critical path; ALGORITHM; SEARCH;
D O I
10.1016/j.swevo.2024.101753
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As the production environment evolves, distributed manufacturing exhibits heterogeneous characteristics, including diverse machines, workers, and production processes. This paper examines a distributed heterogeneous flexible job shop scheduling problem (DHFJSP) with varying processing times. A mixed integer linear programming (MILP) model of the DHFJSP is formulated with the objective of minimizing the makespan. To solve the DHFJSP, we propose a deep Q network-aided automatic design of a variable neighborhood search algorithm (DQN-VNS). By analyzing schedules, sixty-one types of scheduling features are extracted. These features, along with six shaking strategies, are used as states and actions. A DHFJSP environment simulator is developed to train the deep Q network. The well-trained DQN then generates the shaking procedure for VNS. Additionally, a greedy initialization method is proposed to enhance the quality of the initial solution. Seven efficient critical path-based neighborhood structures with three-vector encoding scheme are introduced to improve local search. Numerical experiments on various scales of instances validate the effectiveness of the MILP model and the DQN-VNS algorithm. The results show that the DQN-VNS algorithm achieves an average relative percentage deviation (ARPD) of 3.2%, which represents an approximately 88.45% reduction compared to the best-performing algorithm among the six compared, with an ARPD of 27.7%. This significant reduction in ARPD highlights the superior stability and performance of the proposed DQN-VNS algorithm.
引用
收藏
页数:23
相关论文
共 50 条
  • [31] A novel method for solving dynamic flexible job-shop scheduling problem via DIFFormer and deep reinforcement learning
    Wan, Lanjun
    Cui, Xueyan
    Zhao, Haoxin
    Fu, Long
    Li, Changyun
    COMPUTERS & INDUSTRIAL ENGINEERING, 2024, 198
  • [32] A deep reinforcement learning model for dynamic job-shop scheduling problem with uncertain processing time
    Wu, Xinquan
    Yan, Xuefeng
    Guan, Donghai
    Wei, Mingqiang
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 131
  • [33] Learning to Dispatch for Job Shop Scheduling via Deep Reinforcement Learning
    Zhang, Cong
    Song, Wen
    Cao, Zhiguang
    Zhang, Jie
    Tan, Puay Siew
    Xu, Chi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [34] Dynamic scheduling for multi-objective flexible job shop via deep reinforcement learning
    Yuan, Erdong
    Wang, Liejun
    Song, Shiji
    Cheng, Shuli
    Fan, Wei
    APPLIED SOFT COMPUTING, 2025, 171
  • [35] DeepMAG: Deep reinforcement learning with multi-agent graphs for flexible job shop scheduling
    Zhang, Jia-Dong
    He, Zhixiang
    Chan, Wing -Ho
    Chow, Chi -Yin
    KNOWLEDGE-BASED SYSTEMS, 2023, 259
  • [36] Distributed heterogeneous flexible job-shop scheduling problem considering automated guided vehicle transportation via improved deep Q network
    Yuan, Minghai
    Lu, Songwei
    Zheng, Liang
    Yu, Qi
    Pei, Fengque
    Gu, Wenbin
    SWARM AND EVOLUTIONARY COMPUTATION, 2025, 94
  • [37] Distributional reinforcement learning with the independent learners for flexible job shop scheduling problem with high variability
    Oh, Seung Heon
    Cho, Young In
    Woo, Jong Hun
    JOURNAL OF COMPUTATIONAL DESIGN AND ENGINEERING, 2022, 9 (04) : 1157 - 1174
  • [38] A Reinforcement Learning Approach for Flexible Job Shop Scheduling Problem With Crane Transportation and Setup Times
    Du, Yu
    Li, Junqing
    Li, Chengdong
    Duan, Peiyong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (04) : 5695 - 5709
  • [39] Concurrent scheduling of jobs and AGVs in a flexible job shop system: a parallel hybrid PSO-GA meta-heuristic
    Amirteimoori, Arash
    Kia, Reza
    FLEXIBLE SERVICES AND MANUFACTURING JOURNAL, 2023, 35 (03) : 727 - 753
  • [40] A novel method-based reinforcement learning with deep temporal difference network for flexible double shop scheduling problem
    Wang, Xiao
    Zhong, Peisi
    Liu, Mei
    Zhang, Chao
    Yang, Shihao
    SCIENTIFIC REPORTS, 2024, 14 (01)