Toward Efficient Ensemble Learning with Structure Constraints: Convergent Algorithms and Applications

被引:2
作者
Lin, Shao-Bo [1 ]
Tang, Shaojie [2 ]
Wang, Yao [1 ]
Wang, Di [1 ]
机构
[1] Xi An Jiao Tong Univ, Ctr Intelligent Decis Making & Machine Learning, Sch Management, Xian 710049, Shanxi, Peoples R China
[2] Univ Texas Dallas, Naveen Jindal Sch Management, Richardson, TX 75083 USA
基金
中国国家自然科学基金;
关键词
ensemble learning; boosting; learning theory; convergence; GREEDY APPROXIMATION; BOOSTING ALGORITHMS; CLASSIFIERS; REGRESSION; GRADIENT; RATES; REGULARIZATION; CONSISTENCY; NETWORKS;
D O I
10.1287/ijoc.2022.1224
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Ensemble learning methods, such as boosting, focus on producing a strong classifier based on numerous weak classifiers. In this paper, we develop a novel ensemble learning method called rescaled boosting with truncation (ReBooT) for binary classification by combining well-known rescaling and regularization ideas in boosting. Theoretically, we present some sufficient conditions for the convergence of ReBooT, derive an almost optimal numerical convergence rate, and deduce fast-learning rates in the framework of statistical learning theory. Experimentally, we conduct both toy simulations and four real-world data runs to show the power of ReBooT. Our results show that, compared with the existing boosting algorithms, ReBooT possesses better learning performance and interpretability in terms of solid theoretical guarantees, perfect structure constraints, and good prediction performance.
引用
收藏
页码:3096 / 3116
页数:21
相关论文
共 66 条
[1]  
[Anonymous], 1997, Interior Point Algorithms: Theory and Analysis
[2]   An L2-Boosting Algorithm for Estimation of a Regression Function [J].
Bagirov, Adil M. ;
Clausen, Conny ;
Kohler, Michael .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2010, 56 (03) :1417-1429
[3]   Approximation and learning by greedy algorithms [J].
Barron, Andrew R. ;
Cohen, Albert ;
Dahmen, Wolfgang ;
DeVore, Ronald A. .
ANNALS OF STATISTICS, 2008, 36 (01) :64-94
[4]  
Bartlett PL, 2007, J MACH LEARN RES, V8, P2347
[5]  
Bickel PJ, 2006, J MACH LEARN RES, V7, P705
[6]   Convergence rates of Kernel Conjugate Gradient for random design regression [J].
Blanchard, Gilles ;
Kraemer, Nicole .
ANALYSIS AND APPLICATIONS, 2016, 14 (06) :763-794
[7]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[8]   Boosting algorithms: Regularization, prediction and model fitting [J].
Buehlmann, Peter ;
Hothorn, Torsten .
STATISTICAL SCIENCE, 2007, 22 (04) :477-505
[9]  
Caponnetto A, 2007, FOUND COMPUT MATH, V7, P331, DOI 10.1007/S10208-006-0196-8
[10]   Decision-tree-based knowledge discovery: Single-vs. multi-decision-tree induction [J].
Chang, Namsik ;
Sheng, Olivia R. Liu .
INFORMS JOURNAL ON COMPUTING, 2008, 20 (01) :46-54