Resource management;
Multi-access edge computing;
Optimization;
Internet of Things;
Dynamic scheduling;
Costs;
Cloud computing;
Processor scheduling;
Markov decision processes;
Deep reinforcement learning;
Deep reinforcement learning (DRL);
Markov decision processes (MDPs);
mobile-edge computing;
resource allocation;
task migration;
INTERNET;
IOT;
MEC;
D O I:
10.1109/JIOT.2025.3555503
中图分类号:
TP [自动化技术、计算机技术];
学科分类号:
0812 ;
摘要:
Task migration and resource allocation are essential to integrate available resources for improving the efficiency of mobile-edge computing to support various computation-intensive and delay-sensitive Internet of Things applications, which cooperative optimization has yet been well addressed to achieve desirable performance. In this article, a deep reinforcement learning (DRL)-based adaptive cooperative optimization strategy is presented to fill this gap. Task migration and resource allocation are jointly employed to adaptively integrate available resources within cooperative edge nodes for processing various randomly offloaded tasks. The policy optimization to minimize the system cost and task dropout rates is formulated as Markov decision processes. A DRL algorithm that enhances deep deterministic policy gradient with a dual experience pool is proposed to jointly optimize the task migration and resource allocation in unknown stochastic application environments. Simulation experiments have been conducted to evaluate the performance of the presented strategy, and the results illustrate that it increases system rewards by 17.9%-61.5% and reduces task dropout rate by 5.2%-31.3% comparing with benchmarks.