READYS: A Reinforcement Learning Based Strategy for Heterogeneous Dynamic Scheduling

被引:15
作者
Grinsztajn, Nathan [1 ]
Beaumont, Olivier [2 ]
Jeannot, Emmanuel [2 ]
Preux, Philippe [1 ]
机构
[1] Univ Lille, INRIA, CNRS UMR CRIStAL 9189, Lille, France
[2] Univ Bordeaux, Inria Bordeaux Sud Ouest, Bordeaux, France
来源
2021 IEEE INTERNATIONAL CONFERENCE ON CLUSTER COMPUTING (CLUSTER 2021) | 2021年
关键词
CHOLESKY FACTORIZATION;
D O I
10.1109/Cluster48925.2021.00031
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose READYS, a reinforcement learning algorithm for the dynamic scheduling of computations modeled as a Directed Acyclic Graph (DAGs). Our goal is to develop a scheduling algorithm in which allocation and scheduling decisions are made at runtime, based on the state of the system, as performed in runtime systems such as StarPU or ParSEC. Reinforcement Learning is a natural candidate to achieve this task, since its general principle is to build step by step a strategy that, given the state of the system (the state of the resources and a view of the ready tasks and their successors in our case), makes a decision to optimize a global criterion. Moreover, the use of Reinforcement Learning is natural in a context where the duration of tasks (and communications) is stochastic. We propose READYS that combines Graph Convolutional Networks (GCN) with an Actor-Critic Algorithm (A2C): it builds an adaptive representation of the scheduling problem on the fly and learns a scheduling strategy, aiming at minimizing the makespan. A crucial point is that READYS builds a general scheduling strategy which is neither limited to only one specific application or task graph nor one particular problem size, and that can be used to schedule any DAG. We focus on different types of task graphs originating from linear algebra factorization kernels (CHOLESKY, LU, QR) and we consider heterogeneous platforms made of a few CPUs and GPUs. We first propose to analyze the performance of READYS when learning is performed on a given (platform, kernel, problem size) combination. Using simulations, we show that the scheduling agent obtains performances very similar or even superior to algorithms from the literature, and that it is especially powerful when the scheduling environment contains a lot of uncertainty. We additionally demonstrate that our agent exhibits very promising generalization capabilities. To the best of our knowledge, this is the first paper which shows that reinforcement learning can really be used for dynamic DAG scheduling on heterogeneous resources.
引用
收藏
页码:70 / 81
页数:12
相关论文
共 49 条
  • [1] Abe K., 2019, ARXIV190511623
  • [2] Addanki Ravichandra, 2019, ADV NEUR IN
  • [3] Agullo E., 2011, 2011 9th IEEE/ACS International Conference on Computer Systems and Applications (AICCSA), P217, DOI 10.1109/AICCSA.2011.6126599
  • [4] Agullo E., 2011, Proceedings of the 25th IEEE International Parallel & Distributed Processing Symposium (IPDPS 2011), P932, DOI 10.1109/IPDPS.2011.90
  • [5] Agullo E., 2010, S APPL ACC HIGH PERF
  • [6] Are Static Schedules so Bad ? A Case Study on Cholesky Factorization
    Agullo, Emmanuel
    Beaumont, Olivier
    Eyraud-Dubois, Lionel
    Kumar, Suraj
    [J]. 2016 IEEE 30TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2016), 2016, : 1021 - 1030
  • [7] [Anonymous], 2018, Machine Learning for Combinatorial Optimization: a Methodological Tour d'Horizon
  • [8] [Anonymous], 1979, Computers and Intractablity: A Guide to the Theory of NP-Completeness
  • [9] [Anonymous], 2004, Handbook of Scheduling: Algorithms, Models, and Performance Analysis
  • [10] StarPU: a unified platform for task scheduling on heterogeneous multicore architectures
    Augonnet, Cedric
    Thibault, Samuel
    Namyst, Raymond
    Wacrenier, Pierre-Andre
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2011, 23 (02) : 187 - 198