A boosting ensemble learning based hybrid light gradient boosting machine and extreme gradient boosting model for predicting house prices

被引:40
作者
Sibindi, Racheal [1 ,4 ]
Mwangi, Ronald Waweru [2 ]
Waititu, Anthony Gichuhi [3 ]
机构
[1] Pan African Univ Inst Basic Sci, Dept Math Technol & Innovatc, Nairobi, Kenya
[2] Jomo Kenyatta Univ Agr & Technol, Sch Comp & Informat Technol, Dept Comp, Nairobi, Kenya
[3] Jomo Kenyatta Univ Agr & Technol, Stat & Actuarial Sci Dept, Nairobi, Kenya
[4] Pan African Univ, Inst Basic Sci, Dept Math Technol & Innovat, Nairobi, Kenya
关键词
boosting ensemble learning; light gradient boosting machine; extreme gradient boosting;
D O I
10.1002/eng2.12599
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The implementation of tree-ensemble models has become increasingly essential in solving classification and prediction problems. Boosting ensemble techniques have been widely used as individual machine learning algorithms in predicting house prices. One of the techniques is LGBM algorithm that employs leaf wise growth strategy, reduces loss and improves accuracy during training which results in overfitting. However, XGBoost algorithm uses level wise growth strategy which takes time to compute resulting in higher computation time. Nevertheless, XGBoost has a regularization parameter, implements column sampling and weight reduction on new trees which combats overfitting. This study focuses on developing a hybrid LGBM and XGBoost model in order to prevent overfitting through minimizing variance whilst improving accuracy. Bayesian hyperparameter optimization technique is implemented on the base learners in order to find the best combination of hyperparameters. This resulted in reduced variance (overfitting) in the hybrid model since the regularization parameter values were optimized. The hybrid model is compared to LGBM, XGBoost, Adaboost and GBM algorithms to evaluate its performance in giving accurate house price predictions using MSE, MAE and MAPE evaluation metrics. The hybrid LGBM and XGBoost model outperformed the other models with MSE, MAE and MAPE of 0.193, 0.285, and 0.156 respectively.
引用
收藏
页数:19
相关论文
共 38 条
[1]  
Abou Omar K. B., 2018, PREPRINT
[2]   Enhancing the prediction of student performance based on the machine learning XGBoost algorithm [J].
Asselman, Amal ;
Khaldi, Mohamed ;
Aammou, Souhaib .
INTERACTIVE LEARNING ENVIRONMENTS, 2023, 31 (06) :3360-3379
[3]   A New Approach for Prediction of Solar Radiation with Using Ensemble Learning Algorithm [J].
Basaran, Kivanc ;
Ozcift, Akin ;
Kilinc, Deniz .
ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2019, 44 (08) :7159-7171
[4]  
Buhlmann Peter, 2012, HDB COMPUTATIONAL ST, P985, DOI DOI 10.1007/978-3-642-21551-3_33
[5]   Degradation Trend Prediction of Pumped Storage Unit Based on MIC-LGBM and VMD-GRU Combined Model [J].
Chen, Peng ;
Deng, Yumin ;
Zhang, Xuegui ;
Ma, Li ;
Yan, Yaoliang ;
Wu, Yifan ;
Li, Chaoshun .
ENERGIES, 2022, 15 (02)
[6]  
Chen WN, 2022, SCI REP-UK, V12, DOI 10.1038/s41598-021-04732-4
[7]   Comparison of machine learning models to provide preliminary forecasts of real estate prices [J].
Chou, Jui-Sheng ;
Fleshman, Dillon-Brandon ;
Dinh-Nhat Truong .
JOURNAL OF HOUSING AND THE BUILT ENVIRONMENT, 2022, 37 (04) :2079-2114
[8]   A stacking-based ensemble learning method for earthquake casualty prediction [J].
Cui, Shaoze ;
Yin, Yunqiang ;
Wang, Dujuan ;
Li, Zhiwu ;
Wang, Yanzhang .
APPLIED SOFT COMPUTING, 2021, 101 (101)
[9]   Modeling minimum miscibility pressure of pure/impure CO2-crude oil systems using adaptive boosting support vector regression: Application to gas injection processes [J].
Dargahi-Zarandi, Atefeh ;
Hemmati-Sarapardeh, Abdolhossein ;
Shateri, Mohammadhadi ;
Menad, Nait Amar ;
Ahmadi, Mohammad .
JOURNAL OF PETROLEUM SCIENCE AND ENGINEERING, 2020, 184
[10]  
Ekundayo I., 2020, Master's Thesis