A reinforcement learning based approach for a multiple-load carrier scheduling problem

被引:28
作者
Chen, Ci [1 ]
Xia, Beixin [1 ,2 ]
Zhou, Bing-hai [3 ]
Xi, Lifeng [1 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Ind Engn & Logist Management, Shanghai 200240, Peoples R China
[2] Shanghai Univ, Sch Mech Engn & Automat, Shanghai 200444, Peoples R China
[3] Shanghai Tongji Univ, Dept Ind Engn, Shanghai, Peoples R China
基金
美国国家科学基金会;
关键词
Materials handling; Multi-criteria decision making; Reinforcement learning; Multiple-load carrier scheduling; Look-ahead scheduling; DISPATCHING RULES; ALGORITHM; SELECTION; DESIGN; AGVS;
D O I
10.1007/s10845-013-0852-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper studies the problem of scheduling a multiple-load carrier which is used to deliver parts to line-side buffers of a general assembly (GA) line. In order to maximize the reward of the GA line, both the throughput of the GA line and the material handling distance are considered as scheduling criteria. After formulating the scheduling problem as a reinforcement learning (RL) problem by defining state features, actions and the reward function, we develop a Q() RL algorithm based scheduling approach. To improve performance, forecasted information such as quantities of parts required in a look-ahead horizon is used when we define state features and actions in formulation. Other than applying traditional material handling request generating policy, we use a look-ahead based request generating policy with which material handling requests are generated based not only on current buffer information but also on future part requirement information. Moreover, by utilizing a heuristic dispatching algorithm, the approach is able to handle future requests as well as existing ones. To evaluate the performance of the approach, we conduct simulation experiments to compare the proposed approach with other approaches. Numerical results demonstrate that the policies obtained by the RL approach outperform other approaches.
引用
收藏
页码:1233 / 1245
页数:13
相关论文
共 35 条
[1]  
[Anonymous], P AMER CONTR CONF
[2]   Particle swarm optimization algorithm for a vehicle routing problem with heterogeneous fleet, mixed backhauls, and time windows [J].
Belmecheri, Farah ;
Prins, Christian ;
Yalaoui, Farouk ;
Amodeo, Lionel .
JOURNAL OF INTELLIGENT MANUFACTURING, 2013, 24 (04) :775-789
[3]   Evaluation of automatic guided vehicle systems [J].
Berman, Sigal ;
Schechtman, Edna ;
Edan, Yael .
ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2009, 25 (03) :522-528
[4]   A multiple-criteria real-time scheduling approach for multiple-load carriers subject to LIFO loading constraints [J].
Chen, Ci ;
Xi, Li-feng ;
Zhou, Bing-hai ;
Zhou, Shen-shen .
INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2011, 49 (16) :4787-4806
[5]  
Ci Chen, 2010, Proceedings 2010 Sixth International Conference on Natural Computation (ICNC 2010), P3768, DOI 10.1109/ICNC.2010.5583198
[6]   Testing and classifying vehicle dispatching rules in three real-world settings [J].
de Koster, RBM ;
Le-Anh, T ;
van der Meer, JR .
JOURNAL OF OPERATIONS MANAGEMENT, 2004, 22 (04) :369-386
[7]  
Duda Richard O, 1973, Pattern classification and scene analysis, V3
[8]   CONTROL OF MULTIPRODUCT BULK SERVICE DIFFUSION/OXIDATION PROCESSES [J].
FOWLER, JW ;
HOGG, GL ;
PHILLIPS, DT .
IIE TRANSACTIONS, 1992, 24 (04) :84-96
[9]   Distributed policy search reinforcement learning for job-shop scheduling tasks [J].
Gabel, Thomas ;
Riedmiller, Martin .
INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2012, 50 (01) :41-61
[10]   Dispatching multi-load AGVs in highly automated seaport container terminals [J].
Grunow, M ;
Günther, HO ;
Lehmann, M .
OR SPECTRUM, 2004, 26 (02) :211-235