Tensor Relational Algebra for Distributed Machine Learning System Design

被引:15
作者
Yuan, Binhang [1 ]
Jankov, Dimitrije [1 ]
Zou, Jia [2 ]
Tang, Yuxin [1 ]
Bourgeois, Daniel [1 ]
Jermaine, Chris [1 ]
机构
[1] Rice Univ, Houston, TX 77251 USA
[2] Arizona State Univ, Tempe, AZ 85287 USA
来源
PROCEEDINGS OF THE VLDB ENDOWMENT | 2021年 / 14卷 / 08期
基金
美国国家科学基金会;
关键词
D O I
10.14778/3457390.3457399
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We consider the question: what is the abstraction that should be implemented by the computational engine of a machine learning system? Current machine learning systems typically push whole tensors through a series of compute kernels such as matrix multiplications or activation functions, where each kernel runs on an AI accelerator (ASIC) such as a GPU. This implementation abstraction provides little built-in support for ML systems to scale past a single machine, or for handling large models with matrices or tensors that do not easily fit into the RAM of an ASIC. In this paper, we present an alternative implementation abstraction called the tensor relational algebra (TRA). The TRA is a set-based algebra based on the relational algebra. Expressions in the TRA operate over binary tensor relations, where keys are multi-dimensional arrays and values are tensors. The TRA is easily executed with high efficiency in a parallel or distributed environment, and amenable to automatic optimization. Our empirical study shows that the optimized TRA-based back-end can significantly outperform alternatives for running ML workflows in distributed clusters.
引用
收藏
页码:1338 / 1350
页数:13
相关论文
共 42 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]   LevelHeaded: A Unified Engine for Business Intelligence and Linear Algebra Querying [J].
Aberger, Christopher R. ;
Lamb, Andrew ;
Olukotun, Kunle ;
Re, Christopher .
2018 IEEE 34TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE), 2018, :449-460
[3]   A three-dimensional approach to parallel matrix multiplication [J].
Agarwal, RC ;
Balle, SM ;
Gustavson, FG ;
Joshi, M ;
Palkar, P .
IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 1995, 39 (05) :575-582
[4]  
Balay S, 2019, PETSC USERS MANUAL
[5]   Expressiveness of Matrix and Tensor Query Languages in terms of ML Operators [J].
Barcelo, Pablo ;
Higuera, Nelson ;
Perez, Jorge ;
Subercaseaux, Bernardo .
PROCEEDINGS OF THE 3RD INTERNATIONAL WORKSHOP ON DATA MANAGEMENT FOR END-TO-END MACHINE LEARNING, DEEM 2019, 2019,
[6]  
Barker B., 2015, WORKSH HIGH PERF COM, V262
[7]  
Baumann P., 1998, SIGMOD Record, V27, P575, DOI 10.1145/276305.276386
[8]  
Brown Paul G., 2010, OVERVIEW SCIDB LARGE, P963, DOI [DOI 10.1145/1807167.1807271, 10.1145/1807167.1807271]
[9]  
Cai Zhuhua, 2013, P 2013 ACM SIGMOD IN, P637, DOI [10.1145/2463676, DOI 10.1145/2463676.2465283]
[10]  
Chen TQ, 2018, PROCEEDINGS OF THE 13TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P579