Developing Real-Time Scheduling Policy by Deep Reinforcement Learning

被引:8
|
作者
Bo, Zitong [1 ,2 ]
Qiao, Ying [1 ]
Leng, Chang [1 ]
Wang, Hongan [1 ]
Guo, Chaoping [1 ]
Zhang, Shaohui [3 ]
机构
[1] Chinese Acad Sci, Inst Software, Beijing Key Lab Human Comp Interact, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] Beijing Natl Speed Skating Oval Operat Co Ltd, Beijing, Peoples R China
来源
2021 IEEE 27TH REAL-TIME AND EMBEDDED TECHNOLOGY AND APPLICATIONS SYMPOSIUM (RTAS 2021) | 2021年
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
real-time scheduling; reinforcement learning; multiprocessor system; deep neural network;
D O I
10.1109/RTAS52030.2021.00019
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Designing scheduling policies for multiprocessor real-time systems is challenging since the multiprocessor scheduling problem is NP-complete. The existing heuristics are customized policies that may achieve poor performance under some specific task loads. Thus, a new design pattern is needed to make the multiprocessor scheduling policies perform well under various task loads. In this paper, we investigate a new real-time scheduling policy based on reinforcement learning. For any given real-time task set, our policy can automatically derive a high performance by online learning. Specifically, we model the real-time scheduling process as a multi-agent cooperative game and propose multi-agent self-cooperative learning that overcomes the curse of dimensionality and credit assignment problems. Simulation results show that our approach can learn high-performance policies for various task/system models.
引用
收藏
页码:131 / 142
页数:12
相关论文
共 50 条
  • [41] Real-Time Lane Configuration with Coordinated Reinforcement Learning
    Gunarathna, Udesh
    Xie, Hairuo
    Tanin, Egemen
    Karunasekara, Shanika
    Borovica-Gajic, Renata
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: APPLIED DATA SCIENCE TRACK, ECML PKDD 2020, PT IV, 2021, 12460 : 291 - 307
  • [42] Evolving population method for real-time reinforcement learning
    Kim, Man-Je
    Kim, Jun Suk
    Ahn, Chang Wook
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 229
  • [43] Multi-Agent Reinforcement Learning for Real-Time Dynamic Production Scheduling in a Robot Assembly Cell
    Johnson, Dazzle
    Chen, Gang
    Lu, Yuqian
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (03) : 7684 - 7691
  • [44] Federated Deep Reinforcement Learning Approach for Real-Time Target Coverage in Wireless Sensor Networks
    Glodek, Corey Thomas
    Siasi, Nazli
    2024 IEEE 21ST INTERNATIONAL CONFERENCE ON MOBILE AD-HOC AND SMART SYSTEMS, MASS 2024, 2024, : 641 - 646
  • [45] Deep Reinforcement Learning-Based Control for Real-Time Hybrid Simulation of Civil Structures
    Nino, Andres Felipe
    Palacio-Betancur, Alejandro
    Miranda-Chiquito, Piedad
    Amaya, Juan David
    Silva, Christian E.
    Gutierrez Soto, Mariantonieta
    Felipe Giraldo, Luis
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2025,
  • [46] Panda: Reinforcement Learning-Based Priority Assignment for Multi-Processor Real-Time Scheduling
    Lee, Hyunsung
    Lee, Jinkyu
    Yeom, Ikjun
    Woo, Honguk
    IEEE ACCESS, 2020, 8 : 185570 - 185583
  • [47] Deep Reinforcement Learning with Uncertain Data for Real-Time Stormwater System Control and Flood Mitigation
    Saliba, Sami M.
    Bowes, Benjamin D.
    Adams, Stephen
    Beling, Peter A.
    Goodall, Jonathan L.
    WATER, 2020, 12 (11) : 1 - 19
  • [48] Safe reinforcement learning method integrating process knowledge for real-time scheduling of gas supply network
    Zhou, Pengwei
    Xu, Zuhua
    Zhu, Xiaoping
    Zhao, Jun
    Song, Chunyue
    Shao, Zhijiang
    INFORMATION SCIENCES, 2023, 633 : 280 - 304
  • [49] Research on an Adaptive Real-Time Scheduling Method of Dynamic Job-Shop Based on Reinforcement Learning
    Zhu, Haihua
    Tao, Shuai
    Gui, Yong
    Cai, Qixiang
    MACHINES, 2022, 10 (11)
  • [50] Towards Real-Time Path Planning through Deep Reinforcement Learning for a UAV in Dynamic Environments
    Chao Yan
    Xiaojia Xiang
    Chang Wang
    Journal of Intelligent & Robotic Systems, 2020, 98 : 297 - 309