An ANN-Guided Multi-Objective Framework for Power-Performance Balancing in HPC Systems

被引:0
作者
Maas, William [1 ]
de Souza, Paulo S. S. [2 ]
Luizelli, Marcelo C. [2 ]
Rossi, Fabio D. [3 ]
Navaux, Philippe O. A. [1 ]
Lorenzon, Arthur F. [1 ]
机构
[1] Univ Fed Rio Grande do Sul, Informat Inst, Porto Alegre, RS, Brazil
[2] UNIPAMPA, Alegrete, RS, Brazil
[3] IFFar, Alegrete, RS, Brazil
来源
PROCEEDINGS OF THE 21ST ACM INTERNATIONAL CONFERENCE ON COMPUTING FRONTIERS 2024, CF 2024 | 2024年
关键词
High-Performance Computing; Power-Performance Efficiency; Artificial Neural Network; SCHEDULING ALGORITHM; EXECUTION; EFFICIENT; DVFS; ENERGY;
D O I
10.1145/3649153.3649185
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Power-performance efficiency has become one of the most critical issues in evolving High-Performance Computing systems (HPC) towards Exaflops. Thread-level parallelism (TLP) exploitation, dynamic voltage and frequency scaling (DVFS), and uncore frequency scaling (UFS) are methods widely applied to better balance the power consumption and performance improvements of parallel applications. However, selecting ideal combinations of these knobs for every application is challenging due to the massive number of possible solutions, as there is no unique combination that delivers at the same time the best performance and the lowest power consumption. Given that, we propose HPC-PPO (power-performance optimizer), a multi-objective optimization strategy driven by an artificial neural network that leverages hardware and software features of parallel applications to predict Pareto-efficient configurations of TLP degree, DVFS, and UFS that optimize the balance between power and performance. When validating HPC-PPO on three multicore processors with twenty-five applications, we show that HPC-PPO can predict combinations very close to the best ones found by an exhaustive search. We also show that the Pareto-efficient configurations predicted by HPC-PPO improve parallel applications' performance by 30.7% while spending 23.9% less power when compared to state-of-the-art strategies.
引用
收藏
页码:138 / 146
页数:9
相关论文
共 34 条
[1]   PARMA: Parallelization-Aware Run-Time Management for Energy-Efficient Many-Core Systems [J].
Al-hayanni, Mohammed A. Noaman ;
Rafiev, Ashur ;
Xia, Fei ;
Shafik, Rishad ;
Romanovsky, Alexander ;
Yakovlev, Alex .
IEEE TRANSACTIONS ON COMPUTERS, 2020, 69 (10) :1507-1518
[2]  
Andr~e E., 2020, working paper
[3]   Multi Objective Optimization of HPC Kernels for Performance, Power, and Energy [J].
Balaprakash, Prasanna ;
Tiwari, Ananta ;
Wild, Stefan M. .
HIGH PERFORMANCE COMPUTING SYSTEMS: PERFORMANCE MODELING, BENCHMARKING AND SIMULATION, 2014, 8551 :239-260
[4]  
Benesty J., 2009, Pearson correlation coefficient, DOI [DOI 10.1007/978-3-642-00296-0, 10.1007/978-3-642-00296-0, DOI 10.1007/978-3-642-00296-0_5]
[5]  
Blake G, 2010, CONF PROC INT SYMP C, P302, DOI 10.1145/1816038.1816000
[6]   Modelling DVFS and UFS for Region-Based Energy Aware Tuning of HPC Applications [J].
Chadha, Mohak ;
Gerndt, Michael .
2019 IEEE 33RD INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2019), 2019, :805-814
[7]   Explicit uncore frequency scaling for energy optimisation policies with EAR in Intel architectures [J].
Corbalan, Julita ;
Vidal, Oriol ;
Alonso, Lluis ;
Aneas, Jordi .
2021 IEEE INTERNATIONAL CONFERENCE ON CLUSTER COMPUTING (CLUSTER 2021), 2021, :572-581
[8]   A Reconfiguration Algorithm for Power-Aware Parallel Applications [J].
De Sensi, Daniele ;
Torquati, Massimo ;
Danelutto, Marco .
ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2016, 13 (04)
[9]   CoScale: Coordinating CPU and Memory System DVFS in Server Systems [J].
Deng, Qingyuan ;
Meisner, David ;
Bhattacharjee, Abhishek ;
Wenisch, Thomas F. ;
Bianchini, Ricardo .
2012 IEEE/ACM 45TH INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO-45), 2012, :143-154
[10]   Towards Energy Budget Control in HPC [J].
Dutot, Pierre-Francois ;
Georgiou, Yiannis ;
Glesser, David ;
Lefevre, Laurent ;
Poquet, Millian ;
Rais, Issam .
2017 17TH IEEE/ACM INTERNATIONAL SYMPOSIUM ON CLUSTER, CLOUD AND GRID COMPUTING (CCGRID), 2017, :381-390