Mille Cheval: a GPU-based in-memory high-performance computing framework for accelerated processing of big-data streams

被引:4
作者
Kumar, Vivek [1 ]
Sharma, Dilip Kumar [2 ]
Mishra, Vinay Kumar [3 ]
机构
[1] Dr APJ Abdul Kalam Tech Univ, Lucknow, Uttar Pradesh, India
[2] GLA Univ, Mathura, UP, India
[3] SRMGPC, Lucknow, Uttar Pradesh, India
关键词
GPGPU; Stream; High-performance computing; HPC; In-memory; Single scan; Parallel; CUDA; Nvidia; Kernel; PARALLEL; TIME; CUDA;
D O I
10.1007/s11227-020-03508-3
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Streams are temporally ordered, rapid changing, ample in volume, and infinite in nature. It is nearly impossible to store the entire data stream due to its large volume and high velocity. In this work, the principle of parallelism is employed to accelerate stream data computing. GPU-based high-performance computing (HPC) framework is proposed for accelerated processing of big-data streams using the in-memory data structure. We have implemented three parallel algorithms to prove the viability of the framework. The contributions of Mille Cheval are: (1) the viability of streaming on accelerators to increase throughput, (2) carefully chosen hash algorithms to achieve low collision rate and high randomness, and (3) memory sketches for approximation. The objective is to leverage the power of a single node using in-memory computing and hybrid computing. HPC does not always require high-end hardware but well-designed algorithms. Achievements of Mille Cheval are: (1) relative error is 1.32 when error rate and overestimate rate are chosen as 0.001 and (2) the host memory space requirement is just 63 MB for 1 terabyte of data. The proposed algorithms are pragmatic. It is evident from experimental results that the framework demonstrates 10X speed-up as compared with CPU implementations and 3X speed-up as compared with GPU implementations.
引用
收藏
页码:6936 / 6960
页数:25
相关论文
共 44 条
[1]  
[Anonymous], 2018, ADV NEURAL INF PROCE
[2]  
[Anonymous], 2013, ANDROID 4 2 APIS AND
[3]   GPU LSM: A Dynamic Dictionary Data Structure for the GPU [J].
Ashkiani, Saman ;
Li, Shengren ;
Farach-Colton, Martin ;
Amenta, Nina ;
Owens, John D. .
2018 32ND IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS), 2018, :430-440
[4]   Why wait? Let us start computing while the data is still on the wire [J].
Bhattacharyya, Shilpi ;
Katramatos, Dimitrios ;
Yoo, Shinjae .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2018, 89 :563-574
[5]   CMSS: Sketching based reliable tracking of large network flows [J].
Cafaro, Massimo ;
Epicoco, Italo ;
Pulimeno, Marco .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2019, 101 :770-784
[6]   mrMoulder: A recommendation-based adaptive parameter tuning approach for big data processing platform [J].
Cai, Lin ;
Qi, Yong ;
Wei, Wei ;
Wu, Jinsong ;
Li, Jingwei .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2019, 93 :570-582
[7]  
Choi HJ, 2019, J SUPERCOMPUT, V1, P1
[8]   Efficiency and productivity for decision making on low-power heterogeneous CPU plus GPU SoCs [J].
Constantinescu, Denisa-Andreea ;
Navarro, Angeles ;
Corbera, Francisco ;
Fernandez-Madrigal, Juan-Antonio ;
Asenjo, Rafael .
JOURNAL OF SUPERCOMPUTING, 2021, 77 (01) :44-65
[9]  
Cormode G., 2011, IEEE software, P64, DOI DOI 10.1109/MS.2011.127
[10]  
Curtis, 2011, ACM SIGCOMM, P254