Transferable multi-objective factory layout planning using simulation-based deep reinforcement learning

被引:0
|
作者
Klar, Matthias [1 ,3 ]
Schworm, Philipp [1 ]
Wu, Xiangqian [1 ]
Simon, Peter [1 ]
Glatt, Moritz [1 ]
Ravani, Bahram [2 ]
Aurich, Jan C. [1 ]
机构
[1] RPTU Kaiserslautern, Inst Mfg Technol & Prod Syst, Kaiserslautern, Germany
[2] Univ Calif Davis, Dept Mech & Aerosp Engn, Davis, CA USA
[3] POB 3049, D-67653 Kaiserslautern, Germany
关键词
Facility layout problem; Reinforcement learning; Multi -objective optimization; Discrete event simulation; Material flow; GENETIC ALGORITHM; DESIGN; OPTIMIZATION; SEARCH;
D O I
10.1016/j.jmsy.2024.04.007
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Factory layout planning aims at finding an optimized layout configuration under consideration of varying influences such as the material flow characteristics. Manual layout planning can be characterized as a complex decision-making process due to a large number of possible placement options. Automated planning approaches aim at reducing the manual planning effort by generating optimized layout variants in the early stages of layout planning. Recent developments have introduced deep Reinforcement Learning (RL) based planning approaches that allow to optimize a layout under consideration of a single optimization criterion. However, within layout planning, multiple partially conflicting planning objectives have to be considered. Such multiple objectives are not considered by existing RL-based approaches. This paper addresses this research gap by presenting a novel deep RL-based layout planning approach that allows consideration of multiple objectives for optimization. Furthermore, existing RL-based planning approaches only consider analytically formulated objectives such as the transportation distance. Consequently, dynamic influences in the material flow are neglected which can result in higher operational costs of the future factory. To address this issue, a discrete event simulation module is developed that allows simulating manufacturing and material flow processes simultaneously for any layout configuration generated by the RL approach. Consequently, the presented approach considers material flow simulation results for multi-objective optimization. To investigate the capabilities of RL-based factory layout planning, different RL architectures are compared based on a simplified application scenario. Throughput time, media supply, and material flow clarity are considered as optimization objectives. The best performing architecture is then applied to an exemplary application scenario and compared with the results obtained by a combined version of the genetic algorithm and tabu search, the non-dominated sorting genetic algorithm, and the optimal solution. Finally, two industrial planning scenarios, one focusing on brownfield and one on greenfield planning, are considered. The results show that the performance of RL compared to meta-heuristics depends on the considered computation time. With time the results generated by the RL approach exceed the quality of the best conventional solution by up to 11%. Finally, the potential of applying transfer learning is investigated for three different application scenarios. It is observed that RL can learn generalized patterns for factory layout planning, which allows to significantly reduce the required training time and can lead to improved solution quality. Thus, the use of pre-trained RL models shows a substantial performance potential for automated factory layout planning which cannot be achieved with conventional automated planning approaches.
引用
收藏
页码:487 / 511
页数:25
相关论文
共 50 条
  • [21] A novel multi-objective optimization based multi-agent deep reinforcement learning approach for microgrid resources planning
    Abid, Md. Shadman
    Apon, Hasan Jamil
    Hossain, Salman
    Ahmed, Ashik
    Ahshan, Razzaqul
    Lipu, M. S. Hossain
    APPLIED ENERGY, 2024, 353
  • [22] Multi-Objective Multi-Satellite Imaging Mission Planning Algorithm for Regional Mapping Based on Deep Reinforcement Learning
    Chen, Yaxin
    Shen, Xin
    Zhang, Guo
    Lu, Zezhong
    REMOTE SENSING, 2023, 15 (16)
  • [23] Self-Learning Multi-Objective Service Coordination Using Deep Reinforcement Learning
    Schneider, Stefan
    Khalili, Ramin
    Manzoor, Adnan
    Qarawlus, Haydar
    Schellenberg, Rafael
    Karl, Holger
    Hecker, Artur
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2021, 18 (03): : 3829 - 3842
  • [24] Multi-objective safe reinforcement learning: the relationship between multi-objective reinforcement learning and safe reinforcement learning
    Horie, Naoto
    Matsui, Tohgoroh
    Moriyama, Koichi
    Mutoh, Atsuko
    Inuzuka, Nobuhiro
    ARTIFICIAL LIFE AND ROBOTICS, 2019, 24 (03) : 352 - 359
  • [25] Multi-objective safe reinforcement learning: the relationship between multi-objective reinforcement learning and safe reinforcement learning
    Naoto Horie
    Tohgoroh Matsui
    Koichi Moriyama
    Atsuko Mutoh
    Nobuhiro Inuzuka
    Artificial Life and Robotics, 2019, 24 : 352 - 359
  • [26] Multi-objective pruning of dense neural networks using deep reinforcement learning
    Hirsch, Lior
    Katz, Gilad
    INFORMATION SCIENCES, 2022, 610 : 381 - 400
  • [27] Investigating the multi-objective optimization of quality and efficiency using deep reinforcement learning
    Wang, Zhenhui
    Lu, Juan
    Chen, Chaoyi
    Ma, Junyan
    Liao, Xiaoping
    APPLIED INTELLIGENCE, 2022, 52 (11) : 12873 - 12887
  • [28] Investigating the multi-objective optimization of quality and efficiency using deep reinforcement learning
    Zhenhui Wang
    Juan Lu
    Chaoyi Chen
    Junyan Ma
    Xiaoping Liao
    Applied Intelligence, 2022, 52 : 12873 - 12887
  • [29] Superconducting quantum computing optimization based on multi-objective deep reinforcement learning
    Liu, Yangting
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [30] Multi-objective Dynamic AGV Scheduling Method Based on Deep Reinforcement Learning
    Wang, Gaoshang
    Zou, Yuanyuan
    Li, Shaoyuan
    2024 14TH ASIAN CONTROL CONFERENCE, ASCC 2024, 2024, : 1450 - 1455