End-to-End Multitarget Flexible Job Shop Scheduling With Deep Reinforcement Learning

被引:0
作者
Wang, Rongkai [1 ]
Jing, Yiyang [1 ]
Gu, Chaojie [1 ]
He, Shibo [1 ]
Chen, Jiming [2 ]
机构
[1] Zhejiang Univ, State Key Lab Ind Control Technol, Hangzhou 310027, Peoples R China
[2] Hangzhou Dianzi Univ, Sch Automat, Hangzhou 310018, Peoples R China
来源
IEEE INTERNET OF THINGS JOURNAL | 2025年 / 12卷 / 04期
基金
中国国家自然科学基金;
关键词
Job shop scheduling; Transportation; Production; Manufacturing; Heuristic algorithms; Energy consumption; Optimal scheduling; Metaheuristics; Dispatching; Computer architecture; Cloud-edge manufacturing paradigm; graph neural network (GNN); multiagent reinforcement learning; multitarget flexible job shop scheduling optimization (MT-F[!text type='JS']JS[!/text]P); ALGORITHM;
D O I
10.1109/JIOT.2024.3485748
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Modeling and solving the flexible job shop scheduling problem (FJSP) is critical for modern manufacturing. However, existing works primarily focus on the time-related makespan target, often neglecting other practical factors, such as transportation. To address this, we formulate a more comprehensive multitarget FJSP that integrates makespan with varied transportation times and the total energy consumption of processing and transportation. The combination of these multiple real-world production targets renders the scheduling problem highly complex and challenging to solve. To overcome this challenge, this article proposes an end-to-end multiagent proximal policy optimization (PPO) approach. First, we represent the scheduling problem as a disjunctive graph (DG) with designed features of subtasks and constructed machine nodes, additionally integrating information of arcs denoted as transportation and standby time, respectively. Next, we use a graph neural network (GNN) to encode features into node embeddings, representing the states at each decision step. Finally, based on the vectorized value function and local critic networks, the PPO algorithm and DG simulation environment iteratively interact to train the policy network. Our extensive experimental results validate the performance of the proposed approach, demonstrating its superiority over the state-of-the-art in terms of high-quality solutions, online computation time, stability, and generalization.
引用
收藏
页码:4420 / 4434
页数:15
相关论文
共 45 条
  • [31] Wang Hai, 2023, 2023 IEEE 39th International Conference on Data Engineering (ICDE), P3522, DOI 10.1109/ICDE55515.2023.00269
  • [32] Multiagent and Bargaining-Game-Based Real-Time Scheduling for Internet of Things-Enabled Flexible Job Shop
    Wang, Jin
    Zhang, Yingfeng
    Liu, Yang
    Wu, Naiqi
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (02): : 2518 - 2531
  • [33] An interoperable and flat Industrial Internet of Things architecture for low latency data collection in manufacturing systems
    Wang, Rongkai
    Gu, Chaojie
    He, Shibo
    Shi, Zhiguo
    Meng, Wenchao
    [J]. JOURNAL OF SYSTEMS ARCHITECTURE, 2022, 129
  • [34] Time-Constrained Actor-Critic Reinforcement Learning for Concurrent Order Dispatch in On-Demand Delivery
    Wang, Shuai
    Guo, Baoshen
    Ding, Yi
    Wang, Guang
    He, Suining
    Zhang, Desheng
    He, Tian
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (08) : 8175 - 8192
  • [35] Smart scheduling of dynamic job shop based on discrete event simulation and deep reinforcement learning
    Wang, Ziqing
    Liao, Wenzhu
    [J]. JOURNAL OF INTELLIGENT MANUFACTURING, 2024, 35 (06) : 2593 - 2610
  • [36] A survey of job shop scheduling problem: The types and models
    Xiong, Hegen
    Shi, Shuangyuan
    Ren, Danni
    Hu, Jinjin
    [J]. COMPUTERS & OPERATIONS RESEARCH, 2022, 142
  • [37] Xu J., 2020, P MACHINE LEARNING R, V1, P10607
  • [38] Xu K., 2019, arXiv, DOI 10.48550/arXiv.1810.00826
  • [39] Yang RZ, 2019, ADV NEUR IN, V32
  • [40] Solving flexible job shop scheduling problems via deep reinforcement learning
    Yuan, Erdong
    Wang, Liejun
    Cheng, Shuli
    Song, Shiji
    Fan, Wei
    Li, Yongming
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2024, 245