Markov Decision Process Design for Imitation of Optimal Task Schedulers

被引:0
|
作者
Rademacher, Paul [1 ]
Wagner, Kevin [2 ]
Smith, Leslie [1 ]
机构
[1] US Naval Res Lab, Navy Ctr Appl Res AI, Washington, DC 20375 USA
[2] US Naval Res Lab, Div Radar, Washington, DC USA
来源
2023 IEEE STATISTICAL SIGNAL PROCESSING WORKSHOP, SSP | 2023年
关键词
Scheduling; imitation learning; Markov decision process; tree search;
D O I
10.1109/SSP53291.2023.10207940
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the generally prohibitive computational requirements of optimal task schedulers, much of the field of task scheduling focuses on designing fast suboptimal algorithms. Since the tree search commonly used by sequencing algorithms such as Branch-and-Bound can naturally be framed as a Markov decision process, designing schedulers using imitation and reinforcement learning is a promising and active area of research. This paper demonstrates how polices can be trained on previously solved scheduling problems and successfully generalize to novel ones. Instead of focusing on policy design, however, this work focuses on designing the Markov decision process observation and reward functions to make learning as effective and efficient as possible. This can be of critical importance when training data is limited or when only simple, fast policies are practical. Various Markov decision process designs are introduced and simulation examples demonstrate the resultant increases in policy performance, even without integration into search algorithms.
引用
收藏
页码:56 / 60
页数:5
相关论文
共 50 条
  • [1] A Markov Decision Process Approach for Optimal Data Backup Scheduling
    Xia, Ruofan
    Machida, Fumio
    Trivedi, Kishor
    2014 44TH ANNUAL IEEE/IFIP INTERNATIONAL CONFERENCE ON DEPENDABLE SYSTEMS AND NETWORKS (DSN), 2014, : 660 - 665
  • [2] Optimal passenger-seeking policies on E-hailing platforms using Markov decision process and imitation learning
    Shou, Zhenyu
    Di, Xuan
    Ye, Jieping
    Zhu, Hongtu
    Zhang, Hua
    Hampshire, Robert
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2020, 111 : 91 - 113
  • [3] Optimal electricity supply bidding by Markov decision process
    Song, HL
    Liu, CC
    Lawarrée, J
    Dahlgren, RW
    IEEE TRANSACTIONS ON POWER SYSTEMS, 2000, 15 (02) : 618 - 624
  • [4] Combinatorial Design of a Stochastic Markov Decision Process
    Dimitrov, Nedialko B.
    Morton, David P.
    OPERATIONS RESEARCH AND CYBER-INFRASTRUCTURE, 2009, : 167 - 193
  • [5] Optimal Replacement Policy of Services Based on Markov Decision Process
    Pillai, Sandhya S.
    Narendra, Nanjangud C.
    2009 IEEE INTERNATIONAL CONFERENCE ON SERVICES COMPUTING, 2009, : 176 - +
  • [6] A Markov Decision Process to Determine Optimal Policies in Moving Target
    Zheng, Jianjun
    Namin, Akbar Siami
    PROCEEDINGS OF THE 2018 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'18), 2018, : 2321 - 2323
  • [7] Design of Opportunistic Routing Based on Markov Decision Process
    Hao, Jun
    Jia, Xinchun
    Han, Zongyuan
    Yang, Bo
    Peng, Dengyong
    PROCEEDINGS OF THE 36TH CHINESE CONTROL CONFERENCE (CCC 2017), 2017, : 8976 - 8981
  • [8] A Markov Decision Process Workflow for Automating Interior Design
    Karan, Ebrahim
    Asgari, Sadegh
    Rashidi, Abbas
    KSCE JOURNAL OF CIVIL ENGINEERING, 2021, 25 (09) : 3199 - 3212
  • [9] A Markov Decision Process Workflow for Automating Interior Design
    Ebrahim Karan
    Sadegh Asgari
    Abbas Rashidi
    KSCE Journal of Civil Engineering, 2021, 25 : 3199 - 3212
  • [10] Enhancing the decision making process for paper mill schedulers
    Murthy, S
    Akkiraju, R
    Goodwin, R
    Keskinocak, P
    Wu, F
    Kumaran, S
    Yeh, J
    Fuhrer, R
    Aggarwal, A
    Sturzenbecker, M
    Jayaraman, R
    Daigle, B
    TAPPI FINISHING & CONVERTING CONFERENCE AND TRADE FAIR, 1999, : 7 - 18