Graph neural networks-based scheduler for production planning problems using reinforcement learning

被引:12
|
作者
Hameed, Mohammed Sharafath Abdul [1 ]
Schwung, Andreas [1 ]
机构
[1] South Westphalia Univ Appl Sci, Dept Automat Technol & Learning Syst, D-59494 Soest, Germany
关键词
Job shop scheduling; Reinforcement learning; Graph neural networks; Distributed optimization; Production planning; Tabu search; Genetic algorithm; PROGRAMMING-MODELS; TABU SEARCH; GO; ALGORITHM; SYSTEMS; SHOGI; CHESS;
D O I
10.1016/j.jmsy.2023.06.005
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Reinforcement learning (RL) is increasingly adopted in job shop scheduling problems (JSSP). But RL for JSSP is usually done using a vectorized representation of machine features as the state space. It has three major problems: (1) the relationship between the machine units and the job sequence is not fully captured, (2) exponential increase in the size of the state space with increasing machines/jobs, and (3) the generalization of the agent to unseen scenarios. This paper presents a novel framework named GraSP-RL, GRAph neural network-based Scheduler for Production planning problems using Reinforcement Learning. It represents JSSP as a graph and trains the RL agent using features extracted using a graph neural network (GNN). While the graph is itself in the non-Euclidean space, the features extracted using the GNNs provide a rich encoding of the current production state in the Euclidean space. At its core is a custom message-passing algorithm applied to the GNN. The node features encoded by the GNN are then used by the RL agent to select the next job. Further, we cast the scheduling problem as a decentralized optimization problem in which the learning agent is assigned to all the production units individually and the agent learns asynchronously from the experience collected on all the other production units. The GraSP-RL is then applied to a complex injection molding production environment with 30 jobs and 4 machines. The task is to minimize the makespan of the production plan. The schedule planned by GraSP-RL is then compared and analyzed with a priority dispatch rule algorithm like first-in-first-out (FIFO) and metaheuristics like tabu search (TS) and genetic algorithm (GA). The proposed GraSP-RL outperforms the FIFO, TS, and GA for the trained task of planning 30 jobs in JSSP. We further test the generalization capability of the trained agent on two different problem classes: Open shop system (OSS) and Reactive JSSP (RJSSP). In these modified problem classes our method produces results better than FIFO and comparable results to TS and GA, without any further training while also providing schedules instantly.
引用
收藏
页码:91 / 102
页数:12
相关论文
共 50 条
  • [1] Efficient Integration of Reinforcement Learning in Graph Neural Networks-Based Recommender Systems
    Sharifbaev, Abdurakhmon
    Mozikov, Mikhail
    Zaynidinov, Hakimjon
    Makarov, Ilya
    IEEE ACCESS, 2024, 12 : 189439 - 189448
  • [2] Counterfactual based reinforcement learning for graph neural networks
    Pham, David
    Zhang, Yongfeng
    ANNALS OF OPERATIONS RESEARCH, 2022,
  • [3] Graph neural networks-based preference learning method for object ranking
    Meng, Zhenhua
    Lin, Rongheng
    Wu, Budan
    INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2024, 167
  • [4] HARL: Hierarchical Adaptive Reinforcement Learning Based Auto Scheduler for Neural Networks
    Zhang, Zining
    He, Bingsheng
    Zhang, Zhenjie
    51ST INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2022, 2022,
  • [5] A neural networks-based approach for strategic planning
    Chien, TW
    Lin, C
    Tan, B
    Lee, WC
    INFORMATION & MANAGEMENT, 1999, 35 (06) : 357 - 364
  • [6] A Learning-Based Scheduler for High Volume Processing in Data Warehouse Using Graph Neural Networks
    Bengre, Vivek
    HoseinyFarahabady, M. Reza
    Pivezhandi, Mohammad
    Zomaya, Albert Y.
    Jannesari, Ali
    PARALLEL AND DISTRIBUTED COMPUTING, APPLICATIONS AND TECHNOLOGIES, PDCAT 2021, 2022, 13148 : 175 - 186
  • [7] Structural entropy-based scheduler for job planning problems using multi-agent reinforcement learning
    Liang, Lixin
    Sun, Shuo
    Hao, Zhifeng
    Yang, Yong
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2025,
  • [8] Reinforcement Learning and Trajectory Planning based on Model Approximation with Neural Networks applied to Transition Problems
    Pritzkoleit, Max
    Knoll, Carsten
    Roebenack, Klaus
    IFAC PAPERSONLINE, 2020, 53 (02): : 1581 - 1587
  • [9] Production Scheduling based on Deep Reinforcement Learning using Graph Convolutional Neural Network
    Seito, Takanari
    Munakata, Satoshi
    ICAART: PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 2, 2020, : 766 - 772
  • [10] A graph neural networks-based deep Q-learning approach for job shop scheduling problems in traffic management
    Liu, Zeyi
    Wang, Yuan
    Liang, Xingxing
    Ma, Yang
    Feng, Yanghe
    Cheng, Guangquan
    Liu, Zhong
    INFORMATION SCIENCES, 2022, 607 : 1211 - 1223