Asymptotically Optimal Lagrangian Priority Policy for Deadline Scheduling With Processing Rate Limits

被引:6
作者
Hao, Liangliang [1 ]
Xu, Yunjian [1 ]
Tong, Lang [2 ]
机构
[1] Chinese Univ Hong Kong, Dept Mech & Automat Engn, Hong Kong 999077, Peoples R China
[2] Cornell Univ, Sch Elect & Comp Engn, Ithaca, NY 14853 USA
基金
美国国家科学基金会;
关键词
Program processors; Processor scheduling; Indexes; Dynamic scheduling; Process control; Electric vehicle charging; Time-varying systems; Deadline scheduling; dynamic programming; electric vehicle (EV) charging; index policy; restless multiarmed bandit (RMAB); INDEX POLICIES; ELECTRIC VEHICLES; ALLOCATION; BOUNDS; BANDIT;
D O I
10.1109/TAC.2021.3049340
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We study the deadline scheduling problem for multiple deferrable jobs that arrive in a random manner and are to be processed before individual deadlines. The processing of the jobs is subject to a time-varying limit on the total processing rate at each stage. We formulate the scheduling problem as a restless multiarmed bandit problem. Relaxing the scheduling problem into multiple independent single-arm scheduling problems, we define the Lagrangian priority value as the greatest tax under which it is optimal to activate the arm. We propose a Lagrangian priority policy that processes jobs in the order of their Lagrangian priority values, and establish its asymptotic optimality as the system scales. Numerical results show that the proposed Lagrangian priority policy achieves 22%-49% higher average reward than the classical Whittle index policy (that does not take into account the processing rate limits).
引用
收藏
页码:236 / 250
页数:15
相关论文
共 40 条
[1]   Relaxations of weakly coupled stochastic dynamic programs [J].
Adelman, Daniel ;
Mersereau, Adam J. .
OPERATIONS RESEARCH, 2008, 56 (03) :712-727
[2]  
Altman E., 1999, Constrained Markov Decision Processes
[3]  
Ayesta Urtzi, 2011, 2011 49th Annual Allerton Conference on Communication, Control, and Computing (Allerton), P377
[4]   Scheduling in a Random Environment: Stability and Asymptotic Optimality [J].
Ayesta, Urtzi ;
Erausquin, Martin ;
Jonckheere, Matthieu ;
Verloop, Ina Maria .
IEEE-ACM TRANSACTIONS ON NETWORKING, 2013, 21 (01) :258-271
[5]   A modeling framework for optimizing the flow-level scheduling with time-varying channels [J].
Ayesta, Urtzi ;
Erausqum, Martin ;
Jacko, Peter .
PERFORMANCE EVALUATION, 2010, 67 (11) :1014-1029
[6]   DYNAMIC PROGRAMMING [J].
BELLMAN, R .
SCIENCE, 1966, 153 (3731) :34-&
[7]  
Bertsekas D.P., 2019, Reinforcement Learning and Optimal Control
[8]   OPTIMAL SHORT-TERM SCHEDULING OF LARGE-SCALE POWER-SYSTEMS [J].
BERTSEKAS, DP ;
LAUER, GS ;
SANDELL, NR ;
POSBERGH, TA .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1983, 28 (01) :1-11
[9]  
Bertsekas DP., 2002, INTRO PROBABILITY, V1
[10]   Index Policies and Performance Bounds for Dynamic Selection Problems [J].
Brown, David B. ;
Smith, James E. .
MANAGEMENT SCIENCE, 2020, 66 (07) :3029-3050