Large Iterative Multitier Ensemble Classifiers for Security of Big Data

被引:27
作者
Abawajy, Jemal H. [1 ]
Kelarev, Andrei [1 ]
Chowdhury, Morshed [1 ]
机构
[1] Deakin Univ, Sch Informat Technol, Burwood, Vic 3125, Australia
关键词
LIME classifiers; ensemble meta classifiers; random forest; big data; MALWARE; CHALLENGE;
D O I
10.1109/TETC.2014.2316510
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper introduces and investigates large iterative multitier ensemble (LIME) classifiers specifically tailored for big data. These classifiers are very large, but are quite easy to generate and use. They can be so large that it makes sense to use them only for big data. They are generated automatically as a result of several iterations in applying ensemble meta classifiers. They incorporate diverse ensemble meta classifiers into several tiers simultaneously and combine them into one automatically generated iterative system so that many ensemble meta classifiers function as integral parts of other ensemble meta classifiers at higher tiers. In this paper, we carry out a comprehensive investigation of the performance of LIME classifiers for a problem concerning security of big data. Our experiments compare LIME classifiers with various base classifiers and standard ordinary ensemble meta classifiers. The results obtained demonstrate that LIME classifiers can significantly increase the accuracy of classifications. LIME classifiers performed better than the base classifiers and standard ensemble meta classifiers.
引用
收藏
页码:352 / 363
页数:12
相关论文
共 69 条
  • [1] Abou-Assaleh T, 2004, P INT COMP SOFTW APP, P41
  • [2] Shape quantization and recognition with randomized trees
    Amit, Y
    Geman, D
    [J]. NEURAL COMPUTATION, 1997, 9 (07) : 1545 - 1588
  • [3] [Anonymous], 2013, E HDB STAT METH
  • [4] [Anonymous], 2013, ENCY BRITANNICA
  • [5] [Anonymous], [No title captured]
  • [6] [Anonymous], 2013, EXPANDED TOP 10 BIG
  • [7] Batten L. M., 2011, Proceedings of the 1st International Conference on Cloud Computing and Services Science. CLOSER 2011, P66
  • [8] An empirical comparison of voting classification algorithms: Bagging, boosting, and variants
    Bauer, E
    Kohavi, R
    [J]. MACHINE LEARNING, 1999, 36 (1-2) : 105 - 139
  • [9] Beliakov Gleb, 2012, Journal of Networks, V7, P935, DOI 10.4304/jnw.7.6.935-945
  • [10] Random forests
    Breiman, L
    [J]. MACHINE LEARNING, 2001, 45 (01) : 5 - 32