Waste stabilization pond modelling using extreme gradient boosting machines

被引:4
作者
Ogarekpe, Nkpa Mba [1 ]
Agunwamba, Jonah C. [2 ]
Tenebe, Imokhai T. [3 ,4 ]
Udodi, Obianuju [5 ]
Chinedu, Ani D. [5 ]
机构
[1] Univ Cross River State, Dept Civil Engn, Calabar, Nigeria
[2] Univ Nigeria, Dept Civil Engn, Nsukka, Nigeria
[3] San Jose State Univ, Mineta Transportat Inst, San Jose, CA 95192 USA
[4] Univ Chicago, Booth Business Sch Chicago, Chicago, IL USA
[5] Univ Nigeria, Ctr Environm Management & Control, UNN, SHELL, Nsukka, Nigeria
关键词
ISHJEWSP; machine learning; modelling; optimization; pond; wastewater; INTEGRATED SOLAR; NEURAL-NETWORK; PERFORMANCE; PREDICTION; DESIGN; REMOVAL; PARAMETERS; REGRESSION; SYSTEMS; BOD;
D O I
10.2166/wpt.2024.277
中图分类号
TV21 [水资源调查与水利规划];
学科分类号
081501 ;
摘要
The integrated solar and hydraulic jump-enhanced waste stabilization pond (ISHJEWSP) has been proposed as a solution to enhance performance of the conventional WSP. Despite the better performance of the ISHJEWSP, there is seemingly no previous study that has deployed machine learning (ML) methods in modelling the ISHJEWSP. This study is aimed at determining the relationships between the ISHJEWSP effluent parameters as well as comparing the performance of extra trees (ET), random forest (RF), decision tree (DT), light gradient boosting machine (LightGBM), gradient boosting (GB), and extreme gradient boosting (XGBoost) methods in predicting the effluent biochemical oxygen demand (BOD5) in the ISHJEWSP. The feature importance technique indicated that the most important were pH, temperature, solar radiation, dissolved oxygen (DO), and total suspended solids. These selected features yielded strong correlations with the dependent variable except DO, which had a moderate correlation. With respect to coefficient of determination and root mean square error (RMSE), the XGBoost performed better than the other models [coefficient of determination (R-2) = 0.807, mean absolute error (MAE) = 4.3453, RMSE = 6.2934, root mean squared logarithmic error (RMSLE) = 0.1096]. Gradient boosting, XGBoost, and RF correspondingly yielded the least MAE, RMSE, and RMSLE of 3.9044, 6.2934, and 0.1059, respectively. The study demonstrates effectiveness of ML in predicting the effluent BOD5 in the ISHJEWSP.
引用
收藏
页码:4572 / 4584
页数:13
相关论文
共 69 条
[1]   Prediction of Student's performance by modelling small dataset size [J].
Abu Zohair, Lubna Mahmoud .
INTERNATIONAL JOURNAL OF EDUCATIONAL TECHNOLOGY IN HIGHER EDUCATION, 2019, 16 (01)
[2]   Effect of tapering on the performance of waste stabilization ponds [J].
Agunwamba, JC .
WATER RESEARCH, 2001, 35 (05) :1191-1200
[3]   Predictive modelling for solar thermal energy systems: A comparison of support vector regression, random forest, extra trees and regression trees [J].
Ahmad, Muhammad Waseem ;
Reynolds, Jonathan ;
Rezgui, Yacine .
JOURNAL OF CLEANER PRODUCTION, 2018, 203 :810-821
[4]   An artificial neural network model and design equations for BOD and COD removal prediction in horizontal subsurface flow constructed wetlands [J].
Akratos, Christos S. ;
Papaspyros, John N. E. ;
Tsihrintzis, Vassilios A. .
CHEMICAL ENGINEERING JOURNAL, 2008, 143 (1-3) :96-110
[5]   Predicting Breast Cancer from Risk Factors Using SVM and Extra-Trees-Based Feature Selection Method [J].
Alfian, Ganjar ;
Syafrudin, Muhammad ;
Fahrurrozi, Imam ;
Fitriyani, Norma Latif ;
Atmaji, Fransiskus Tatas Dwi ;
Widodo, Tri ;
Bahiyah, Nurul ;
Benes, Filip ;
Rhee, Jongtae .
COMPUTERS, 2022, 11 (09)
[6]   A Novel Extra Tree Ensemble Optimized DL Framework (ETEODL) for Early Detection of Diabetes [J].
Arya, Monika ;
Sastry, Hanumat. G. ;
Motwani, Anand ;
Kumar, Sunil ;
Zaguia, Atef .
FRONTIERS IN PUBLIC HEALTH, 2022, 9
[7]  
Ayyadevara VK., 2018, PROMACHINE LEARNING, DOI [DOI 10.1007/978-1-4842-3564-5_6, 10.1007/978-1-4842-3564-5, DOI 10.1007/978-1-4842-3564-5, DOI 10.1007/978-1-4842-3564-56]
[8]   A comparative analysis of gradient boosting algorithms [J].
Bentejac, Candice ;
Csorgo, Anna ;
Martinez-Munoz, Gonzalo .
ARTIFICIAL INTELLIGENCE REVIEW, 2021, 54 (03) :1937-1967
[9]   A random forest guided tour [J].
Biau, Gerard ;
Scornet, Erwan .
TEST, 2016, 25 (02) :197-227
[10]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32