Developing Real-Time Scheduling Policy by Deep Reinforcement Learning

被引:12
作者
Bo, Zitong [1 ,2 ]
Qiao, Ying [1 ]
Leng, Chang [1 ]
Wang, Hongan [1 ]
Guo, Chaoping [1 ]
Zhang, Shaohui [3 ]
机构
[1] Chinese Acad Sci, Inst Software, Beijing Key Lab Human Comp Interact, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Beijing Natl Speed Skating Oval Operat Co Ltd, Beijing, Peoples R China
来源
2021 IEEE 27TH REAL-TIME AND EMBEDDED TECHNOLOGY AND APPLICATIONS SYMPOSIUM (RTAS 2021) | 2021年
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
real-time scheduling; reinforcement learning; multiprocessor system; deep neural network;
D O I
10.1109/RTAS52030.2021.00019
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Designing scheduling policies for multiprocessor real-time systems is challenging since the multiprocessor scheduling problem is NP-complete. The existing heuristics are customized policies that may achieve poor performance under some specific task loads. Thus, a new design pattern is needed to make the multiprocessor scheduling policies perform well under various task loads. In this paper, we investigate a new real-time scheduling policy based on reinforcement learning. For any given real-time task set, our policy can automatically derive a high performance by online learning. Specifically, we model the real-time scheduling process as a multi-agent cooperative game and propose multi-agent self-cooperative learning that overcomes the curse of dimensionality and credit assignment problems. Simulation results show that our approach can learn high-performance policies for various task/system models.
引用
收藏
页码:131 / 142
页数:12
相关论文
共 35 条
[1]   Early-release fair scheduling [J].
Anderson, JH ;
Srinivasan, A .
EUROMICRO RTS 2000: 12TH EUROMICRO CONFERENCE ON REAL-TIME SYSTEMS, PROCEEDINGS, 2000, :35-43
[2]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[3]  
Baruah S. K., 1995, Proceedings 9th International Parallel Processing Symposium (Cat. No.95TH8052), P280, DOI 10.1109/IPPS.1995.395946
[4]  
Baruah SK, 1996, ALGORITHMICA, V15, P600, DOI 10.1007/BF01940883
[5]   NEW STRATEGIES FOR ASSIGNING REAL-TIME TASKS TO MULTIPROCESSOR SYSTEMS [J].
BURCHARD, A ;
LIEBEHERR, J ;
OH, YF ;
SON, SH .
IEEE TRANSACTIONS ON COMPUTERS, 1995, 44 (12) :1429-1442
[6]   A comprehensive survey of multiagent reinforcement learning [J].
Busoniu, Lucian ;
Babuska, Robert ;
De Schutter, Bart .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2008, 38 (02) :156-172
[7]  
Busoniu L, 2010, STUD COMPUT INTELL, V310, P183
[8]  
Cho K., 2014, C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG C EMP METH NAT LANG, P1724
[9]  
Chu X, 2017, ARXIVABS171000336 CO
[10]   Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis [J].
Dai, Angela ;
Qi, Charles Ruizhongtai ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6545-6554