Graph neural networks-based scheduler for production planning problems using reinforcement learning

被引:12
|
作者
Hameed, Mohammed Sharafath Abdul [1 ]
Schwung, Andreas [1 ]
机构
[1] South Westphalia Univ Appl Sci, Dept Automat Technol & Learning Syst, D-59494 Soest, Germany
关键词
Job shop scheduling; Reinforcement learning; Graph neural networks; Distributed optimization; Production planning; Tabu search; Genetic algorithm; PROGRAMMING-MODELS; TABU SEARCH; GO; ALGORITHM; SYSTEMS; SHOGI; CHESS;
D O I
10.1016/j.jmsy.2023.06.005
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Reinforcement learning (RL) is increasingly adopted in job shop scheduling problems (JSSP). But RL for JSSP is usually done using a vectorized representation of machine features as the state space. It has three major problems: (1) the relationship between the machine units and the job sequence is not fully captured, (2) exponential increase in the size of the state space with increasing machines/jobs, and (3) the generalization of the agent to unseen scenarios. This paper presents a novel framework named GraSP-RL, GRAph neural network-based Scheduler for Production planning problems using Reinforcement Learning. It represents JSSP as a graph and trains the RL agent using features extracted using a graph neural network (GNN). While the graph is itself in the non-Euclidean space, the features extracted using the GNNs provide a rich encoding of the current production state in the Euclidean space. At its core is a custom message-passing algorithm applied to the GNN. The node features encoded by the GNN are then used by the RL agent to select the next job. Further, we cast the scheduling problem as a decentralized optimization problem in which the learning agent is assigned to all the production units individually and the agent learns asynchronously from the experience collected on all the other production units. The GraSP-RL is then applied to a complex injection molding production environment with 30 jobs and 4 machines. The task is to minimize the makespan of the production plan. The schedule planned by GraSP-RL is then compared and analyzed with a priority dispatch rule algorithm like first-in-first-out (FIFO) and metaheuristics like tabu search (TS) and genetic algorithm (GA). The proposed GraSP-RL outperforms the FIFO, TS, and GA for the trained task of planning 30 jobs in JSSP. We further test the generalization capability of the trained agent on two different problem classes: Open shop system (OSS) and Reactive JSSP (RJSSP). In these modified problem classes our method produces results better than FIFO and comparable results to TS and GA, without any further training while also providing schedules instantly.
引用
收藏
页码:91 / 102
页数:12
相关论文
共 50 条
  • [21] Reinforcement Learning Enhanced Explainer for Graph Neural Networks
    Shan, Caihua
    Shen, Yifei
    Zhang, Yao
    Li, Xiang
    Li, Dongsheng
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [22] A Deep Reinforcement Learning Heuristic for SAT based on Antagonist Graph Neural Networks
    Fournier, Thomas
    Lallouet, Arnaud
    Cropsal, Telio
    Glorian, Gael
    Papadopoulos, Alexandre
    Petitet, Antoine
    Perez, Guillaume
    Sekar, Suruthy
    Suijlen, Wijnand
    2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 1218 - 1222
  • [23] TD based reinforcement learning using neural networks in control problems with continuous action space
    Lee, JH
    Oh, SY
    Choi, DH
    IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE, 1998, : 2028 - 2033
  • [24] A Reinforcement Learning System with Chaotic Neural Networks-Based Adaptive Hierarchical Memory Structure for Autonomous Robots
    Obayashi, Masanao
    Narita, Kenichiro
    Kuremoto, Takashi
    Kobayashi, Kunikazu
    2008 INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS, VOLS 1-4, 2008, : 80 - 85
  • [25] Graph convolutional neural networks-based assessment of students' collaboration ability
    Lin, Jinjiao
    Gao, Tianqi
    Wen, Yuhua
    Yu, Xianmiao
    You, Bizhen
    Yin, Yanfang
    Zhao, Yanze
    Pu, Haitao
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2022, 34 (28):
  • [26] USING MODULAR NEURAL NETWORKS AND MACHINE LEARNING WITH REINFORCEMENT LEARNING TO SOLVE CLASSIFICATION PROBLEMS
    Leoshchenko, S. D.
    Oliinyk, A. O.
    Subbotin, S. A.
    Kolpakova, T. O.
    RADIO ELECTRONICS COMPUTER SCIENCE CONTROL, 2024, (02) : 71 - 81
  • [27] UNTANGLE: Unlocking Routing and Logic Obfuscation Using Graph Neural Networks-based Link Prediction
    Alrahis, Lilas
    Patnaik, Satwik
    Hanif, Muhammad Abdullah
    Shafique, Muhammad
    Sinanoglu, Ozgur
    2021 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN (ICCAD), 2021,
  • [28] Artificial Neural Networks-Based Machine Learning for Wireless Networks: A Tutorial
    Chen, Mingzhe
    Challita, Ursula
    Saad, Walid
    Yin, Changchuan
    Debbah, Merouane
    IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2019, 21 (04): : 3039 - 3071
  • [29] ReLeS: A Neural Adaptive Multipath Scheduler based on Deep Reinforcement Learning
    Zhang, Han
    Li, Wenzhong
    Gao, Shaohua
    Wang, Xiaoliang
    Ye, Baoliu
    IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2019), 2019, : 1648 - 1656
  • [30] Forward and bidirectional planning based on reinforcement learning and neural networks in a simulated robot
    Baldassarre, G
    ANTICIPATORY BEHAVIOR IN ADAPTIVE LEARNING SYSTEMS: FOUNDATIONS, THEORIES, AND SYSTEMS, 2003, 2684 : 179 - 200