Design patterns of deep reinforcement learning models for job shop scheduling problems

被引:6
作者
Wang, Shiyong [1 ]
Li, Jiaxian [1 ]
Jiao, Qingsong [2 ]
Ma, Fang [3 ]
机构
[1] South China Univ Technol, Sch Mech & Automot Engn, Guangzhou 510640, Peoples R China
[2] South China Univ Technol, Dept Elect Business, Guangzhou 510640, Peoples R China
[3] China Natl Elect Apparat Res Inst Co Ltd, Guangzhou 510300, Peoples R China
基金
国家重点研发计划;
关键词
Production scheduling; Reinforcement learning; Smart manufacturing; Industry; 4.0; OPTIMIZATION;
D O I
10.1007/s10845-024-02454-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Production scheduling has a significant role when optimizing production objectives such as production efficiency, resource utilization, cost control, energy-saving, and emission reduction. Currently, deep reinforcement learning-based production scheduling methods achieve roughly equivalent precision as the widely used meta-heuristic algorithms while exhibiting higher efficiency, along with powerful generalization abilities. Therefore, this new paradigm has drawn much attention and plenty of research results have been reported. By reviewing available deep reinforcement learning models for the job shop scheduling problems, the typical design patterns and pattern combinations of the common components, i.e., agent, environment, state, action, and reward, were identified. Around this essential contribution, the architecture and procedure of training deep reinforcement learning scheduling models and applying resultant scheduling solvers were generalized. Furthermore, the key evaluation indicators were summarized and the promising research areas were outlined. This work surveys several deep reinforcement learning models for a range of production scheduling problems.
引用
收藏
页数:19
相关论文
共 75 条
  • [1] End-to-end on-line rescheduling from Gantt chart images using deep reinforcement learning
    Andres Palombarini, Jorge
    Carlos Martinez, Ernesto
    [J]. INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2022, 60 (14) : 4434 - 4463
  • [2] Constructing a Search Strategy and Searching for Evidence
    Aromataris, Edoardo
    Riitano, Dagmara
    [J]. AMERICAN JOURNAL OF NURSING, 2014, 114 (05) : 49 - 56
  • [3] Deep Reinforcement Learning A brief survey
    Arulkumaran, Kai
    Deisenroth, Marc Peter
    Brundage, Miles
    Bharath, Anil Anthony
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) : 26 - 38
  • [4] Aveyard H., 2016, A post-graduate's guide to doing a literature review in health and social care
  • [5] Bettany-Saltikov J., 2012, How to do a systematic literature review in Nursing: A stepby- step guide
  • [6] Natural actor-critic algorithms
    Bhatnagar, Shalabh
    Sutton, Richard S.
    Ghavamzadeh, Mohammad
    Lee, Mark
    [J]. AUTOMATICA, 2009, 45 (11) : 2471 - 2482
  • [7] Boland A., 2017, DOING SYSTEMATIC REV
  • [8] Deep Reinforcement Learning for Dynamic Flexible Job Shop Scheduling with Random Job Arrival
    Chang, Jingru
    Yu, Dong
    Hu, Yi
    He, Wuwei
    Yu, Haoyu
    [J]. PROCESSES, 2022, 10 (04)
  • [9] A self-learning genetic algorithm based on reinforcement learning for flexible job-shop scheduling problem
    Chen, Ronghua
    Yang, Bo
    Li, Shi
    Wang, Shilong
    [J]. COMPUTERS & INDUSTRIAL ENGINEERING, 2020, 149
  • [10] A Deep Reinforcement Learning Framework Based on an Attention Mechanism and Disjunctive Graph Embedding for the Job-Shop Scheduling Problem
    Chen, Ruiqi
    Li, Wenxin
    Yang, Hongbing
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (02) : 1322 - 1331