A reinforcement learning-based approach for online optimal control of self-adaptive real-time systems

被引:3
作者
Haouari, Bakhta [1 ,2 ,3 ]
Mzid, Rania [1 ,4 ]
Mosbahi, Olfa [3 ]
机构
[1] Univ Tunis El Manar, ISI, 2 Rue Abourraihan Al Bayrouni, Ariana 2080, Tunisia
[2] Univ Carthage, Ctr Urbain Nord, LISI Lab INSAT, BP 676, Tunis 1080, Tunisia
[3] Univ Carthage, Tunisia Polytech Sch, BP 743, La Marsa 2078, Tunisia
[4] Univ Sfax, CES Lab ENIS, BP W3, Sfax 3038, Tunisia
关键词
Embedded real-time systems; Self-adaptation; Scheduling; Online reinforcement learning; Optimal control; Robustness;
D O I
10.1007/s00521-023-08778-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper deals with self-adaptive real-time embedded systems (RTES). A self-adaptive system can operate in different modes. Each mode encodes a set of real-time tasks. To be executed, each task is allocated to a processor (placement) and assigned a priority (scheduling), while respecting timing constraints. An adaptation scenario allows switching between modes by adding, removing, and updating task parameters that must meet related deadlines after adaptation. For such systems, anticipating all operational modes at design time is usually impossible. Online reinforcement learning is increasingly used in the presence of design-time uncertainty. To tackle this problem, we formalize the placement and scheduling problems in self-adaptive RTES as a Markov decision process and propose related algorithms based on Q-learning. Then, we introduce an approach that integrates the proposed algorithms to assist designers in the development of self-adaptive RTES. At the design level, the RL Placement and the RL Scheduler are proposed to process predictable adaptation scenarios. These modules are designed to generate placement and scheduling models for an application while maximizing system extensibility and ensuring real-time feasibility. At the execution level, the RL Adapter is defined to process online adaptations. Indeed, the goal of the RL Adapter agent is to reject the adaptation scenario when feasibility concerns are raised; otherwise, it generates a new feasible placement and scheduling. We apply and simulate the proposed approach to a healthcare robot case study to show its applicability. Performance evaluations are conducted to prove the effectiveness of the proposed approach compared to related works.
引用
收藏
页码:20375 / 20401
页数:27
相关论文
共 64 条
[1]   A novel hybrid multilayer perceptron neural network with improved grey wolf optimizer [J].
Altay, Osman ;
Altay, Elif Varol .
NEURAL COMPUTING & APPLICATIONS, 2023, 35 (01) :529-556
[2]   Deep Reinforcement Learning A brief survey [J].
Arulkumaran, Kai ;
Deisenroth, Marc Peter ;
Brundage, Miles ;
Bharath, Anil Anthony .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) :26-38
[3]   APPLYING NEW SCHEDULING THEORY TO STATIC PRIORITY PREEMPTIVE SCHEDULING [J].
AUDSLEY, N ;
BURNS, A ;
RICHARDSON, M ;
TINDELL, K ;
WELLINGS, AJ .
SOFTWARE ENGINEERING JOURNAL, 1993, 8 (05) :284-292
[4]  
Audsley N.C., 1990, DEADLINE MONOTONIC S
[5]   On priority assignment in fixed priority scheduling [J].
Audsley, NC .
INFORMATION PROCESSING LETTERS, 2001, 79 (01) :39-44
[6]  
Barto A. G., 2021, SIAM Review, V6, P423
[7]   A MARKOVIAN DECISION PROCESS [J].
BELLMAN, R .
JOURNAL OF MATHEMATICS AND MECHANICS, 1957, 6 (05) :679-684
[8]   Multi-objective design exploration approach for Ravenscar real-time systems [J].
Bouaziz, Rahma ;
Lemarchand, Laurent ;
Singhoff, Frank ;
Zalila, Bechir ;
Jmaiel, Mohamed .
REAL-TIME SYSTEMS, 2018, 54 (02) :424-483
[9]   Architecture Exploration of Real-time Systems Based on Multi-Objective Optimization [J].
Bouaziz, Rahma ;
Lemarchand, Laurent ;
Singhoff, Frank ;
Zalila, Bechir ;
Jmaiel, Mohamed .
2015 20TH INTERNATIONAL CONFERENCE ON ENGINEERING OF COMPLEX COMPUTER SYSTEMS (ICECCS), 2015, :1-10
[10]   NEW STRATEGIES FOR ASSIGNING REAL-TIME TASKS TO MULTIPROCESSOR SYSTEMS [J].
BURCHARD, A ;
LIEBEHERR, J ;
OH, YF ;
SON, SH .
IEEE TRANSACTIONS ON COMPUTERS, 1995, 44 (12) :1429-1442