Optimization of hyperparameters and feature selection for random forests and support vector machines by artificial bee colony algorithm

被引:0
作者
Kondo H. [1 ]
Asanuma Y. [2 ]
机构
[1] Department of Intelligent Systems Engineering, College of Engineering, Ibaraki University
[2] Major in Intelligent Systems Engineering, Graduate School of Science and Engineering, Ibaraki University
来源
Transactions of the Japanese Society for Artificial Intelligence | 2019年 / 34卷 / 02期
关键词
Artificial bee colony algorithm; Feature selection; Hyperparameters optimization; Random forests; Support vector machines;
D O I
10.1527/tjsai.G-I36
中图分类号
学科分类号
摘要
Hyperparameters optimization for learning algorithms and feature selection from given data are key issues in machine learning, would greatly affect classification accuracy. Random forests and support vector machines are among some of the most popular learning algorithms. Random forests that have relatively few hyperparameters, can perform more accurate classification by optimizing these parameters without requirement of feature selection. Same as random forests, support vector machines also have a few hyperparameters. However, whether or not to perform feature selection at the same time as optimization of these parameters greatly affects classification accuracy. Usually, grid search method is used to optimize hyperparameters. However, since this search method is performed on predetermined grids, the detailed optimization cannot be realized. In this paper, we thus introduce an artificial bee colony (ABC) algorithm to optimize hyperparameters and to perform more accurate feature selection. ABC algorithm is one of the swarm intelligence algorithms used to solve optimization problems which is inspired by the foraging behaviour of the honey bees. Using KDD Cup 1999 Data that is a benchmark of network intrusion detection classification, experimental results demonstrate the effectiveness of our method. The proposed method is superior in classification accuracies to existing methods for the same data, where swarm intelligence is used to hyperparameters optimization and feature selection. Our method also shows better performance than classification accuracies of random forests and SVM that are learned using default parameters values provided by scikit-learn, an open source machine learning library for Python. © 2019, Japanese Society for Artificial Intelligence. All rights reserved.
引用
收藏
相关论文
共 22 条
  • [1] Bosman P.A.N., (2017)
  • [2] Breiman L., Random forests, Machine Learning, 45, 1, pp. 5-32, (2001)
  • [3] Cortes C., Vapnik V., Support-vector networks, Machine Learning, 20, 3, pp. 273-297, (1995)
  • [4] Criminisi A., Shotton J., Konukoglu E., Decision Forests for Classification, Regression, Density Estimation, Manifold Learning and Semi-Supervised Learning, (2011)
  • [5] Eberhart R., Kennedy J., A New Optimizer Using Particle Swarm Theory, pp. 39-43, (1995)
  • [6] Eberhart R.C., Shi Y., Kennedy J., Swarm Intelligence, (2001)
  • [7] Holland J.H., Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence, (1992)
  • [8] Karaboga D., An Idea Based on Honey Bee Swarm for Numerical Optimization, 200, (2005)
  • [9] Karaboga D., Basturk B., On the performance of artificial bee colony (ABC) algorithm, Applied Soft Computing, 8, 1, pp. 687-697, (2008)
  • [10] Karaboga D., Akay B., A comparative study of artificial bee colony algorithm, Applied Mathematics and Computation, 214, 1, pp. 108-132, (2009)