Balancing Performance and Energy Consumption of Bagging Ensembles for the Classification of Data Streams in Edge Computing

被引:2
作者
Cassales, Guilherme [1 ]
Gomes, Heitor Murilo [2 ]
Bifet, Albert [1 ]
Pfahringer, Bernhard [1 ]
Senger, Hermes [3 ]
机构
[1] Univ Waikato, Dept Comp Sci, Hamilton 3240, New Zealand
[2] Victoria Univ Wellington, Sch Engn & Comp Sci, Wellington 6140, New Zealand
[3] Univ Fed Sao Carlos, Dept Comp Sci, BR-13565905 Sao Carlos, Brazil
来源
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT | 2023年 / 20卷 / 03期
基金
巴西圣保罗研究基金会;
关键词
Edge computing; machine learning; data stream classification; ensembles; energy consumption;
D O I
10.1109/TNSM.2022.3226505
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, the Edge Computing (EC) paradigm has emerged as an enabling factor for developing technologies like the Internet of Things (IoT) and 5G networks, bridging the gap between Cloud Computing services and end-users, supporting low latency, mobility, and location awareness to delay-sensitive applications. An increasing number of solutions in EC have employed machine learning (ML) methods to perform data classification and other information processing tasks on continuous and evolving data streams. Usually, such solutions have to cope with vast amounts of data that come as data streams while balancing energy consumption, latency, and the predictive performance of the algorithms. Ensemble methods achieve remarkable predictive performance when applied to evolving data streams due to several models and the possibility of selective resets. This work investigates a strategy that introduces short intervals to defer the processing of mini-batches. Well balanced, our strategy can improve the performance (i.e., delay, throughput) and reduce the energy consumption of bagging ensembles to classify data streams. The experimental evaluation involved six state-of-art ensemble algorithms (OzaBag, OzaBag Adaptive Size Hoeffding Tree, Online Bagging ADWIN, Leveraging Bagging, Adaptive RandomForest, and Streaming Random Patches) applying five widely used machine learning benchmark datasets with varied characteristics on three computer platforms. As a result, our strategy can significantly reduce energy consumption in 96% of the experimental scenarios evaluated. Despite the trade-offs, it is possible to balance them to avoid significant loss in predictive performance.
引用
收藏
页码:3038 / 3054
页数:17
相关论文
共 37 条
[31]   Data Stream Clustering: A Survey [J].
Silva, Jonathan A. ;
Faria, Elaine R. ;
Barros, Rodrigo C. ;
Hruschka, Eduardo R. ;
de Carvalho, Andre C. P. L. F. ;
Gama, Joao .
ACM COMPUTING SURVEYS, 2013, 46 (01)
[32]  
Snowdon D. C., 2005, P 2005 WORKSH POW AW, V31, P34
[33]   Evaluating the Four-Way Performance Trade-Off for Stream Classification [J].
Turrisi da Costa, Victor G. ;
Santana, Everton Jose ;
Lopes, Jessica F. ;
Barbon, Sylvio, Jr. .
GREEN, PERVASIVE, AND CLOUD COMPUTING, GPC 2019, 2019, 11484 :3-17
[34]   Strict Very Fast Decision Tree: A memory conservative algorithm for data stream mining [J].
Turrisi da Costa, Victor Guilherme ;
Ponce de Leon Ferreira de Carvalho, Andre Carlos ;
Barbon Junior, Sylvio .
PATTERN RECOGNITION LETTERS, 2018, 116 :22-28
[35]   kNN-STUFF: kNN STreaming Unit for Fpgas [J].
Vieira, Joao ;
Duarte, Rui P. ;
Neto, Horacio C. .
IEEE ACCESS, 2019, 7 :170864-170877
[36]   The New Frontier of Smart Grids [J].
Yu, Xinghuo ;
Cecati, Carlo ;
Dillon, Tharam ;
Simoes, M. Godoy .
IEEE INDUSTRIAL ELECTRONICS MAGAZINE, 2011, 5 (03) :49-63
[37]   A Relational Theory of Locality [J].
Yuan, Liang ;
Ding, Chen ;
Smith, Wesley ;
Denning, Peter ;
Zhang, Yunquan .
ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2019, 16 (03)