Evolving Gradient Boost: A Pruning Scheme Based on Loss Improvement Ratio for Learning Under Concept Drift

被引:17
作者
Wang, Kun [1 ,2 ]
Lu, Jie [1 ]
Liu, Anjin [1 ]
Zhang, Guangquan [1 ]
Xiong, Li [2 ]
机构
[1] Univ Technol Sydney, Australia Artificial Intelligence Inst, Fac Engn & Informat Technol, Sydney, NSW 2007, Australia
[2] Shanghai Univ, Sch Management, Shanghai 200444, Peoples R China
基金
澳大利亚研究理事会;
关键词
Bagging; Boosting; Heuristic algorithms; Data models; Adaptation models; Training; Australia; Concept drift; data stream; decision tree; ensemble learning; ONLINE; CLASSIFIERS;
D O I
10.1109/TCYB.2021.3109796
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In nonstationary environments, data distributions can change over time. This phenomenon is known as concept drift, and the related models need to adapt if they are to remain accurate. With gradient boosting (GB) ensemble models, selecting which weak learners to keep/prune to maintain model accuracy under concept drift is nontrivial research. Unlike existing models such as AdaBoost, which can directly compare weak learners' performance by their accuracy (a metric between [0, 1]), in GB, weak learners' performance is measured with different scales. To address the performance measurement scaling issue, we propose a novel criterion to evaluate weak learners in GB models, called the loss improvement ratio (LIR). Based on LIR, we develop two pruning strategies: 1) naive pruning (NP), which simply deletes all learners with increasing loss and 2) statistical pruning (SP), which removes learners if their loss increase meets a significance threshold. We also devise a scheme to dynamically switch between NP and SP to achieve the best performance. We implement the scheme as a concept drift learning algorithm, called evolving gradient boost (LIR-eGB). On average, LIR-eGB delivered the best performance against state-of-the-art methods on both stationary and nonstationary data.
引用
收藏
页码:2110 / 2123
页数:14
相关论文
共 56 条
[1]  
Ashfahani A, 2019, Data Min, P666
[2]   Paired Learners for Concept Drift [J].
Bach, Stephen H. ;
Maloof, Marcus A. .
ICDM 2008: EIGHTH IEEE INTERNATIONAL CONFERENCE ON DATA MINING, PROCEEDINGS, 2008, :23-32
[3]  
Bifet A, 2010, LECT NOTES ARTIF INT, V6321, P135, DOI 10.1007/978-3-642-15880-3_15
[4]  
Bifet A, 2010, J MACH LEARN RES, V11, P1601
[5]  
Bifet A, 2009, KDD-09: 15TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, P139
[6]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[7]   Reacting to Different Types of Concept Drift: The Accuracy Updated Ensemble Algorithm [J].
Brzezinski, Dariusz ;
Stefanowski, Jerzy .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (01) :81-94
[8]  
Brzezinski D, 2011, LECT NOTES ARTIF INT, V6679, P155, DOI 10.1007/978-3-642-21222-2_19
[9]   Kappa Updated Ensemble for drifting data stream mining [J].
Cano, Alberto ;
Krawczyk, Bartosz .
MACHINE LEARNING, 2020, 109 (01) :175-218
[10]   A COEFFICIENT OF AGREEMENT FOR NOMINAL SCALES [J].
COHEN, J .
EDUCATIONAL AND PSYCHOLOGICAL MEASUREMENT, 1960, 20 (01) :37-46