Weighted double deep Q-network based reinforcement learning for bi-objective multi-workflow scheduling in the cloud

被引:23
作者
Li, Huifang [1 ]
Huang, Jianghang [1 ]
Wang, Binyang [1 ]
Fan, Yushun [2 ]
机构
[1] Beijing Inst Technol, State Key Lab Intelligent Control & Decis Complex, Beijing 100081, Peoples R China
[2] Tsinghua Univ, Dept Automat, Beijing 100084, Peoples R China
来源
CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS | 2022年 / 25卷 / 02期
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Multi-objective workflow scheduling; Weighted double deep Q-networks; Cloud computing; OPTIMIZATION; ALGORITHM; GAME;
D O I
10.1007/s10586-021-03454-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As a promising distributed paradigm, cloud computing provides a cost-effective deploying environment for hosting scientific applications due to its provisioning elastic, heterogeneous resources in a pay-per-use model. More and more applications modeled as workflows are being moved to the cloud, and time and cost become important for workflow execution. However, scheduling workflows is still a challenge due to their large-scale and complexity, as well as the cloud's dynamic characteristics and different quotations. In this work, we propose a Weighted Double Deep Q-Network-based Reinforcement Learning algorithm (WDDQN-RL) for scheduling multiple workflows to obtain near-optimal solutions in a relatively short time with both makespan and cost minimized. Specifically, we first introduce a dynamic coefficient-based adaptive balancing method into WDDQN to improve the accuracy of the target value estimation by making a trade-off between Deep Q-Network (DQN) overestimation and Double Deep Q-Network (DDQN) underestimation. Second, pointer network-based agents and a two-level scheduling strategy are designed, where pointer networks are used to process a variable candidate task set in the first-level and one selected task is fed to agents in the second-level for allocating resources. Third, we present a dynamic sensing mechanism by adjusting the model's attention to each individual objective for increasing the diversity of solutions while guaranteeing their quality. Experimental results show that our algorithm outperforms the benchmarking approaches in various indicators.
引用
收藏
页码:751 / 768
页数:18
相关论文
共 50 条
  • [1] Amazon, 2021, AM EC2 ON DEM PRIC
  • [2] Multiagent Reinforcement Learning: Rollout and Policy Iteration
    Bertsekas, Dimitri
    [J]. IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2021, 8 (02) : 249 - 272
  • [3] Bharathi S, 2008, 2008 THIRD WORKSHOP ON WORKFLOWS IN SUPPORT OF LARGE-SCALE SCIENCE (WORKS 2008), P11
  • [4] Budget aware scheduling algorithm for workflow applications in IaaS clouds
    Chakravarthi, K.
    Shyamala, L.
    Vaidehi, V.
    [J]. CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2020, 23 (04): : 3405 - 3419
  • [5] Handling multiple objectives with particle swarm optimization
    Coello, CAC
    Pulido, GT
    Lechuga, MS
    [J]. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2004, 8 (03) : 256 - 279
  • [6] Cui D, 2015, INT S COMPUTATIONAL, P305
  • [7] A fast and elitist multiobjective genetic algorithm: NSGA-II
    Deb, K
    Pratap, A
    Agarwal, S
    Meyarivan, T
    [J]. IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2002, 6 (02) : 182 - 197
  • [8] Task scheduling based on deep reinforcement learning in a cloud manufacturing environment
    Dong, Tingting
    Xue, Fei
    Xiao, Chuangbai
    Li, Juntao
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2020, 32 (11)
  • [9] Multi-objective workflow scheduling in Amazon EC2
    Durillo, Juan J.
    Prodan, Radu
    [J]. CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2014, 17 (02): : 169 - 189
  • [10] An improved task scheduling algorithm for scientific workflow in cloud computing environment
    Geng, Xiaozhong
    Mao, Yingshuang
    Xiong, Mingyuan
    Liu, Yang
    [J]. CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2019, 22 (Suppl 3): : S7539 - S7548