Deep reinforcement learning driven trajectory-based meta-heuristic for distributed heterogeneous flexible job shop scheduling problem

被引:1
|
作者
Zhang, Qichen [1 ]
Shao, Weishi [1 ,3 ,4 ]
Shao, Zhongshi [2 ]
Pi, Dechang [4 ]
Gao, Jiaquan [1 ,3 ]
机构
[1] Nanjing Normal Univ, Sch Comp & Elect Informat, Sch Artificial Intelligence, Nanjing, Peoples R China
[2] Shaanxi Normal Univ, Sch Comp Sci, Xian, Peoples R China
[3] Minist Educ, Key Lab Numer Simulat Large Scale Complex Syst, Beijing, Peoples R China
[4] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing, Peoples R China
基金
中国博士后科学基金;
关键词
Distributed heterogeneous flexible job shop; scheduling problem; Deep Q network; Variable neighborhood search; Makespan; Critical path; ALGORITHM; SEARCH;
D O I
10.1016/j.swevo.2024.101753
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As the production environment evolves, distributed manufacturing exhibits heterogeneous characteristics, including diverse machines, workers, and production processes. This paper examines a distributed heterogeneous flexible job shop scheduling problem (DHFJSP) with varying processing times. A mixed integer linear programming (MILP) model of the DHFJSP is formulated with the objective of minimizing the makespan. To solve the DHFJSP, we propose a deep Q network-aided automatic design of a variable neighborhood search algorithm (DQN-VNS). By analyzing schedules, sixty-one types of scheduling features are extracted. These features, along with six shaking strategies, are used as states and actions. A DHFJSP environment simulator is developed to train the deep Q network. The well-trained DQN then generates the shaking procedure for VNS. Additionally, a greedy initialization method is proposed to enhance the quality of the initial solution. Seven efficient critical path-based neighborhood structures with three-vector encoding scheme are introduced to improve local search. Numerical experiments on various scales of instances validate the effectiveness of the MILP model and the DQN-VNS algorithm. The results show that the DQN-VNS algorithm achieves an average relative percentage deviation (ARPD) of 3.2%, which represents an approximately 88.45% reduction compared to the best-performing algorithm among the six compared, with an ARPD of 27.7%. This significant reduction in ARPD highlights the superior stability and performance of the proposed DQN-VNS algorithm.
引用
收藏
页数:23
相关论文
共 50 条
  • [41] GA based heuristic for the open job shop scheduling problem
    Senthilkumar, P.
    Shahabudeen, P.
    INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2006, 30 (3-4): : 297 - 301
  • [42] GA based heuristic for the open job shop scheduling problem
    P. Senthilkumar
    P. Shahabudeen
    The International Journal of Advanced Manufacturing Technology, 2006, 30 : 297 - 301
  • [43] Concurrent scheduling of jobs and AGVs in a flexible job shop system: a parallel hybrid PSO-GA meta-heuristic
    Arash Amirteimoori
    Reza Kia
    Flexible Services and Manufacturing Journal, 2023, 35 : 727 - 753
  • [44] A dynamic flexible job shop scheduling method based on collaborative agent reinforcement learning
    Shao, Changshun
    Yu, Zhenglin
    Ding, Hongchang
    Cao, Guohua
    Ding, Kaifang
    Duan, Jingsong
    FLEXIBLE SERVICES AND MANUFACTURING JOURNAL, 2024,
  • [45] An intelligent hybrid meta-heuristic for solving a case of no-wait two-stage flexible flow shop scheduling problem with unrelated parallel machines
    Rabiee, M.
    Rad, Reza Sadeghi
    Mazinani, M.
    Shafaei, R.
    INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2014, 71 (5-8): : 1229 - 1245
  • [46] Diverse policy generation for the flexible job-shop scheduling problem via deep reinforcement learning with a novel graph representation
    Echeverria, Imanol
    Murua, Maialen
    Santana, Roberto
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 139
  • [47] A Deep Reinforcement Advantage Actor-Critic-Based Co-Evolution Algorithm for Energy-Aware Distributed Heterogeneous Flexible Job Shop Scheduling
    Xu, Hua
    Tao, Juntai
    Huang, Lingxiang
    Zhang, Chenjie
    Zheng, Jianlu
    PROCESSES, 2025, 13 (01)
  • [48] A self-learning discrete salp swarm algorithm based on deep reinforcement learning for dynamic job shop scheduling problem
    Gu, Yiming
    Chen, Ming
    Wang, Liang
    APPLIED INTELLIGENCE, 2023, 53 (15) : 18925 - 18958
  • [49] Flexible Job Shop Scheduling via Dual Attention Network-Based Reinforcement Learning
    Wang, Runqing
    Wang, Gang
    Sun, Jian
    Deng, Fang
    Chen, Jie
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3091 - 3102
  • [50] Residual Scheduling: A New Reinforcement Learning Approach to Solving Job Shop Scheduling Problem
    Ho, Kuo-Hao
    Cheng, Jui-Yu
    Wu, Ji-Han
    Chiang, Fan
    Chen, Yen-Chi
    Wu, Yuan-Yu
    Wu, I-Chen
    IEEE ACCESS, 2024, 12 : 14703 - 14718