Dynamic Dispatching for Large-Scale Heterogeneous Fleet via Multi-agent Deep Reinforcement Learning

被引:28
作者
Zhang, Chi [1 ]
Odonkor, Philip [2 ]
Zheng, Shuai [1 ]
Khorasgani, Hamed [1 ]
Serita, Susumu [1 ]
Gupta, Chetan [1 ]
Wang, Haiyan [1 ]
机构
[1] Hitachi Amer Ltd, Ind AI Lab, Santa Clara, CA 95054 USA
[2] Stevens Inst Technol, Hoboken, NJ 07030 USA
来源
2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA) | 2020年
关键词
Dispatching; Reinforcement Learning; Mining;
D O I
10.1109/BigData50022.2020.9378191
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dynamic dispatching is one of the core problems for operation optimization in traditional industries such as mining, as it is about how to smartly allocate the right resources to the right place at the right time. Conventionally, the industry relies on heuristics or even human intuitions which are often short-sighted and sub-optimal solutions. Leveraging the power of AI and Internet of Things (IoT), data-driven automation is reshaping this area. However, facing its own challenges such as large-scale and heterogenous trucks running in a highly dynamic environment, it can barely adopt methods developed in other domains (e.g., ride-sharing). In this paper, we propose a novel Deep Reinforcement Learning approach to solve the dynamic dispatching problem in mining. We first develop an event-based mining simulator with parameters calibrated in real mines. Then we propose an experience-sharing Deep Q Network with a novel abstract state/action representation to learn memories from heterogeneous agents altogether and realizes learning in a centralized way. We demonstrate that the proposed methods significantly outperform the most widely adopted approaches in the industry by 5.56% in terms of productivity. The proposed approach has great potential in a broader range of industries (e.g., manufacturing, logistics) which have a large-scale of heterogenous equipment working in a highly dynamic environment, as a general framework for dynamic resource allocation.
引用
收藏
页码:1436 / 1441
页数:6
相关论文
共 50 条
[21]   Multi-agent deep reinforcement learning via double attention and adaptive entropy [J].
Wu, Pei-Liang ;
Yuan, Xu-Dong ;
Mao, Bing-Yi ;
Chen, Wen-Bai ;
Gao, Guo-Wei .
Kongzhi Lilun Yu Yingyong/Control Theory and Applications, 2024, 41 (10) :1930-1936
[22]   Hierarchical reinforcement learning via dynamic subspace search for multi-agent planning [J].
Ma, Aaron ;
Ouimet, Michael ;
Cortes, Jorge .
AUTONOMOUS ROBOTS, 2020, 44 (3-4) :485-503
[23]   Hierarchical reinforcement learning via dynamic subspace search for multi-agent planning [J].
Aaron Ma ;
Michael Ouimet ;
Jorge Cortés .
Autonomous Robots, 2020, 44 :485-503
[24]   Dynamic Scholarly Collaborator Recommendation via Competitive Multi-Agent Reinforcement Learning [J].
Zhang, Yang ;
Zhang, Chenwei ;
Liu, Xiaozhong .
PROCEEDINGS OF THE ELEVENTH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS'17), 2017, :331-335
[25]   A review of cooperative multi-agent deep reinforcement learning [J].
Afshin Oroojlooy ;
Davood Hajinezhad .
Applied Intelligence, 2023, 53 :13677-13722
[26]   Experience Selection in Multi-Agent Deep Reinforcement Learning [J].
Wang, Yishen ;
Zhang, Zongzhang .
2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2019), 2019, :864-870
[27]   Multi-Agent Deep Reinforcement Learning with Emergent Communication [J].
Simoes, David ;
Lau, Nuno ;
Reis, Luis Paulo .
2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
[28]   A review of cooperative multi-agent deep reinforcement learning [J].
Oroojlooy, Afshin ;
Hajinezhad, Davood .
APPLIED INTELLIGENCE, 2023, 53 (11) :13677-13722
[29]   Teaching on a Budget in Multi-Agent Deep Reinforcement Learning [J].
Ilhan, Ercument ;
Gow, Jeremy ;
Perez-Liebana, Diego .
2019 IEEE CONFERENCE ON GAMES (COG), 2019,
[30]   Robust Optimal Formation Control of Heterogeneous Multi-Agent System via Reinforcement Learning [J].
Lin, Wei ;
Zhao, Wanbing ;
Liu, Hao .
IEEE ACCESS, 2020, 8 (08) :218424-218432