Intelligent Scheduling Method for Bulk Cargo Terminal Loading Process Based on Deep Reinforcement Learning

被引:10
作者
Li, Changan [1 ,2 ]
Wu, Sirui [3 ]
Li, Zhan [3 ,4 ]
Zhang, Yuxiao [3 ]
Zhang, Lijie [1 ]
Gomes, Luis [5 ]
机构
[1] Yanshan Univ, Key Lab Adv Forging & Stamping Technol & Sci, Minist Educ China, Qinhuangdao 066004, Hebei, Peoples R China
[2] Chnenergy Tianjin Port Co Ltd, Tianjin 300450, Peoples R China
[3] Harbin Inst Technol, Res Inst Intelligent Control & Syst, Harbin 150001, Peoples R China
[4] Ningbo Inst Intelligent Equipment Technol Co Ltd, Ningbo 315201, Peoples R China
[5] NOVA Univ Lisbon, NOVA Sch Sci & Technol, Ctr Technol & Syst, P-2829516 Monte De Caparica, Portugal
基金
中国国家自然科学基金;
关键词
bulk cargo loading; MDP model; deep reinforcement learning; intelligent scheduling; BERTH ALLOCATION PROBLEM; OPTIMIZATION; SYSTEM;
D O I
10.3390/electronics11091390
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sea freight is one of the most important ways for the transportation and distribution of coal and other bulk cargo. This paper proposes a method for optimizing the scheduling efficiency of the bulk cargo loading process based on deep reinforcement learning. The process includes a large number of states and possible choices that need to be taken into account, which are currently performed by skillful scheduling engineers on site. In terms of modeling, we extracted important information based on actual working data of the terminal to form the state space of the model. The yard information and the demand information of the ship are also considered. The scheduling output of each convey path from the yard to the cabin is the action of the agent. To avoid conflicts of occupying one machine at same time, certain restrictions are placed on whether the action can be executed. Based on Double DQN, an improved deep reinforcement learning method is proposed with a fully connected network structure and selected action sets according to the value of the network and the occupancy status of environment. To make the network converge more quickly, an improved new epsilon-greedy exploration strategy is also proposed, which uses different exploration rates for completely random selection and feasible random selection of actions. After training, an improved scheduling result is obtained when the tasks arrive randomly and the yard state is random. An important contribution of this paper is to integrate the useful features of the working time of the bulk cargo terminal into a state set, divide the scheduling process into discrete actions, and then reduce the scheduling problem into simple inputs and outputs. Another major contribution of this article is the design of a reinforcement learning algorithm for the bulk cargo terminal scheduling problem, and the training efficiency of the proposed algorithm is improved, which provides a practical example for solving bulk cargo terminal scheduling problems using reinforcement learning.
引用
收藏
页数:18
相关论文
共 50 条
[31]   A Distributed Intelligent Lighting Control System Based on Deep Reinforcement Learning [J].
Fang, Peixin ;
Wang, Ming ;
Li, Jingzheng ;
Zhao, Qianchuan ;
Zheng, Xuehan ;
Gao, He .
APPLIED SCIENCES-BASEL, 2023, 13 (16)
[32]   Task scheduling based on deep reinforcement learning in a cloud manufacturing environment [J].
Dong, Tingting ;
Xue, Fei ;
Xiao, Chuangbai ;
Li, Juntao .
CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2020, 32 (11)
[33]   Deep Reinforcement Learning-based Scheduling for Roadside Communication Networks [J].
Atallah, Rihal ;
Assi, Chadi ;
Khahhaz, Maurice .
2017 15TH INTERNATIONAL SYMPOSIUM ON MODELING AND OPTIMIZATION IN MOBILE, AD HOC, AND WIRELESS NETWORKS (WIOPT), 2017,
[34]   Intelligent Control of Manipulator Based on Deep Reinforcement Learning [J].
Zhou, Jiangtao ;
Zheng, Hua ;
Zhao, Dongzhu ;
Chen, Yingxue .
2021 12TH INTERNATIONAL CONFERENCE ON MECHANICAL AND AEROSPACE ENGINEERING (ICMAE), 2021, :275-279
[35]   Beam Hopping Scheduling Based on Deep Reinforcement Learning [J].
Deng, Huimin ;
Ying, Kai ;
Gui, Lin .
2023 INTERNATIONAL CONFERENCE ON FUTURE COMMUNICATIONS AND NETWORKS, FCN, 2023,
[36]   A Deep Reinforcement Learning Method for SAR Earth Observation Satellite Scheduling Problem [J].
Chen, Yingguo ;
Chun, Jie ;
Chen, Ming ;
Liu, Xiaolu ;
Chen, Yingwu .
2024 10TH INTERNATIONAL CONFERENCE ON BIG DATA AND INFORMATION ANALYTICS, BIGDIA 2024, 2024, :687-692
[37]   Deep Reinforcement Learning for Channel Traffic Scheduling in Dry Bulk Export Terminals [J].
Wang, Wenyuan ;
Ding, Ao ;
Cao, Zhen ;
Peng, Yun ;
Liu, Huakun ;
Xu, Xinglu .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (11) :17547-17561
[38]   Scheduling intelligent charging robots for electric vehicle: A deep reinforcement learning approach [J].
Ding, Yi ;
Deng, Ming ;
Ke, Ginger Y. ;
Shen, Yingjun ;
Zhang, Lianmin .
TRANSPORTATION RESEARCH PART E-LOGISTICS AND TRANSPORTATION REVIEW, 2025, 200
[39]   Deep Reinforcement Learning Based Charging Scheduling for Household Electric Vehicles in Active Distribution Network [J].
Qi, Taoyi ;
Ye, Chengjin ;
Zhao, Yuming ;
Li, Lingyang ;
Ding, Yi .
JOURNAL OF MODERN POWER SYSTEMS AND CLEAN ENERGY, 2023, 11 (06) :1890-1901
[40]   A collaborative scheduling method for cloud computing heterogeneous workflows based on deep reinforcement learning [J].
Chen, Genxin ;
Qi, Jin ;
Sun, Ying ;
Hu, Xiaoxuan ;
Dong, Zhenjiang ;
Sun, Yanfei .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2023, 141 :284-297