Model-Free Control for Distributed Stream Data Processing using Deep Reinforcement Learning

被引:54
作者
Li, Teng [1 ]
Xu, Zhiyuan [1 ]
Tang, Jian [1 ]
Wang, Yanzhi [1 ]
机构
[1] Syracuse Univ, Dept Elect Engn & Comp Sci, Syracuse, NY 13244 USA
来源
PROCEEDINGS OF THE VLDB ENDOWMENT | 2018年 / 11卷 / 06期
关键词
SYSTEMS;
D O I
10.14778/3184470.3184474
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we focus on general-purpose Distributed Stream Data Processing Systems (DSDPSs), which deal with processing of unbounded streams of continuous data at scale distributedly in real or near-real time. A fundamental problem in a DSDPS is the scheduling problem (i.e., assigning workload to workers/machines) with the objective of minimizing average end-to-end tuple processing time. A widely-used solution is to distribute workload evenly over machines in the cluster in a round-robin manner, which is obviously not efficient due to lack of consideration for communication delay. Model-based approaches (such as queueing theory) do not work well either due to the high complexity of the system environment. We aim to develop a novel model-free approach that can learn to well control a DSDPS from its experience rather than accurate and mathematically solvable system models, just as a human learns a skill (such as cooking, driving, swimming, etc). Specifically, we, for the first time, propose to leverage emerging Deep Reinforcement Learning (DRL) for enabling model-free control in DSDPSs; and present design, implementation and evaluation of a novel and highly effective DRL-based control framework, which minimizes average end-to-end tuple processing time by jointly learning the system environment via collecting very limited runtime statistics data and making decisions under the guidance of powerful Deep Neural Networks (DNNs). To validate and evaluate the proposed framework, we implemented it based on a widely-used DSDPS, Apache Storm, and tested it with three representative applications: continuous queries, log stream processing and word count (stream version). Extensive experimental results show 1) Compared to Storm's default scheduler and the state-of-the-art model-based method, the proposed framework reduces average tuple processing by 33.5% and 14.0% respectively on average. 2) The proposed framework can quickly reach a good scheduling solution during online learning, which justifies its practicability for online control in DSDPSs.
引用
收藏
页码:705 / 718
页数:14
相关论文
共 27 条
[1]   MillWheel: Fault-Tolerant Stream Processing at Internet Scale [J].
Akidau, Tyler ;
Balikov, Alex ;
Bekiroglu, Kaya ;
Chernyak, Slava ;
Haberman, Josh ;
Lax, Reuven ;
McVeety, Sam ;
Mills, Daniel ;
Nordstrom, Paul ;
Whittle, Sam .
PROCEEDINGS OF THE VLDB ENDOWMENT, 2013, 6 (11) :1033-1044
[2]   M3: Stream Processing on Main-Memory MapReduce [J].
Aly, Ahmed M. ;
Sallam, Asmaa ;
Gnanasekaran, Bala M. ;
Long-Van Nguyen-Dinh ;
Aref, Walid G. ;
Ouzzani, Mourad ;
Ghafoor, Arif .
2012 IEEE 28TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE), 2012, :1253-1256
[3]  
Anderson Q., 2013, Storm real-time processing cookbook: Efficiently process unbounded streams of data in real time
[4]  
Aniello L., P ACM DEBS 2013
[5]  
[Anonymous], P C EMP METH NAT LAN
[6]  
[Anonymous], 2015, Reinforcement Learning: An Introduction
[7]  
Backman N., P MAPREDUCE 12
[8]  
Bellavista P, 2014, INT CONF UTIL CLOUD, P363, DOI 10.1109/UCC.2014.46
[9]  
Boyd L., 2004, CONVEX OPTIMIZATION
[10]   FlumeJava']Java: Easy, Efficient Data-Parallel Pipelines [J].
Chambers, Craig ;
Raniwala, Ashish ;
Perry, Frances ;
Adams, Stephen ;
Henry, Robert R. ;
Bradshaw, Robert ;
Weizenbaum, Nathan .
PLDI '10: PROCEEDINGS OF THE 2010 ACM SIGPLAN CONFERENCE ON PROGRAMMING LANGUAGE DESIGN AND IMPLEMENTATION, 2010, :363-375