HeAT - a Distributed and GPU-accelerated Tensor Framework for Data Analytics

被引:5
作者
Goetz, Markus [4 ]
Debus, Charlotte [1 ]
Coquelin, Daniel [2 ,3 ,4 ]
Krajsek, Kai [3 ]
Comito, Claudia [3 ]
Knechtges, Philipp [1 ]
Hagemeier, Bjorn [3 ]
Tarnawa, Michael [3 ]
Hanselmann, Simon [4 ]
Siggel, Martin [1 ]
Basermann, Achim [1 ]
Streit, Achim [4 ]
机构
[1] German Aerosp Ctr, Inst Software Technol SC, Cologne, Germany
[2] Forschungszentrum Julich, Inst Rio & Geosci Agrosphere IBG 3, Julich, Germany
[3] Forschungszentrum Julich FZJ, Julich Supercomp Ctr JSC, Julich, Germany
[4] Karlsruhe Inst Technol KIT, Steinbuch Ctr Comp SCC, Karlsruhe, Germany
来源
2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA) | 2020年
关键词
IleAT; Tensor Framework; High-performance Computing; PyTorch; NumPy; Message Passing Interface; CPU; Rig Data Analytics; Machine Learning; Dask; Model Parallelism; Parallel Application Frameworks; MODEL;
D O I
10.1109/BigData50022.2020.9378050
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To cope with the rapid growth in available data, the efficiency of data analysis and machine learning libraries has recently received increased attention. Although great advancements have been made in traditional array-based computations, most are limited by the resources available on a single computation node. Consequently, novel approaches must be made to exploit distributed resources, e.g. distributed memory architectures. To this end, we introduce IleAT, an array-based numerical programming framework for large-scale parallel processing with an easy-to-use NumPy-like API. HeAT utilizes PyTorch as a node-local eager execution engine and distributes the workload on arbitrarily large high-performance computing systems via MPI. It provides both low-level array computations, as well as assorted higher-level algorithms. With HeAT, it is possible for a NumPy user to take full advantage of their available resources, significantly I owering the bartier to distributed data analysis. When compared to similar frameworks, HeAT achieves speedups of up to two orders of magnitude.
引用
收藏
页码:276 / 287
页数:12
相关论文
共 38 条
[1]  
Abadi Martin, 2016, Proceedings of OSDI '16: 12th USENIX Symposium on Operating Systems Design and Implementation. OSDI '16, P265
[2]  
[Anonymous], 2018, CoRR abs/1802.05799
[3]   Searching for exotic particles in high-energy physics with deep learning [J].
Baldi, P. ;
Sadowski, P. ;
Whiteson, D. .
NATURE COMMUNICATIONS, 2014, 5
[4]   Legate NumPy: Accelerated and Distributed Array Computing [J].
Bauer, Michael ;
Garland, Michael .
PROCEEDINGS OF SC19: THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2019,
[5]   Untitled [J].
Bennett, John .
EUROPEAN EARLY CHILDHOOD EDUCATION RESEARCH JOURNAL, 2009, 17 (01) :1-6
[6]  
Bradbury James, 2018, JAX: Composable transformations of Python+NumPy programs
[7]   Benchmarking Harp-DAAL: High Performance Hadoop on KNL Clusters [J].
Chen, Langshi ;
Peng, Bo ;
Zhang, Bingjing ;
Liu, Tony ;
Zou, Yiming ;
Jiang, Lei ;
Henschel, Robert ;
Stewart, Craig ;
Zhang, Zhang ;
Mccallum, Emily ;
Tom, Zahniser ;
Jon, Omer ;
Qiu, Judy .
2017 IEEE 10TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING (CLOUD), 2017, :82-89
[8]  
Chen T., 2015, MXNET FLEXIBLE EFFIC
[9]  
Collette Andrew., 2013, Python and HDF5
[10]   The Cityscapes Dataset for Semantic Urban Scene Understanding [J].
Cordts, Marius ;
Omran, Mohamed ;
Ramos, Sebastian ;
Rehfeld, Timo ;
Enzweiler, Markus ;
Benenson, Rodrigo ;
Franke, Uwe ;
Roth, Stefan ;
Schiele, Bernt .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3213-3223