Cancer Classification with a Cost-Sensitive Naive Bayes Stacking Ensemble

被引:22
|
作者
Xiong, Yueling [1 ]
Ye, Mingquan [1 ]
Wu, Changrong [2 ]
机构
[1] Wannan Med Coll, Sch Med Informat, Wuhu 241002, Peoples R China
[2] Anhui Normal Univ, Sch Comp & Informat, Wuhu 241002, Peoples R China
基金
中国国家自然科学基金;
关键词
PARTICLE SWARM OPTIMIZATION; FEATURE-SELECTION; NEURAL-NETWORK; MACHINE;
D O I
10.1155/2021/5556992
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Ensemble learning combines multiple learners to perform combinatorial learning, which has advantages of good flexibility and higher generalization performance. To achieve higher quality cancer classification, in this study, the fast correlation-based feature selection (FCBF) method was used to preprocess the data to eliminate irrelevant and redundant features. Then, the classification was carried out in the stacking ensemble learner. A library for support vector machine (LIBSVM), K-nearest neighbor (KNN), decision tree C4.5 (C4.5), and random forest (RF) were used as the primary learners of the stacking ensemble. Given the imbalanced characteristics of cancer gene expression data, the embedding cost-sensitive naive Bayes was used as the metalearner of the stacking ensemble, which was represented as CSNB stacking. The proposed CSNB stacking method was applied to nine cancer datasets to further verify the classification performance of the model. Compared with other classification methods, such as single classifier algorithms and ensemble algorithms, the experimental results showed the effectiveness and robustness of the proposed method in processing different types of cancer data. This method may therefore help guide cancer diagnosis and research.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Classification of Gold-Bearing Particles Using Visual Cues and Cost-Sensitive Machine Learning
    Horrocks, Tom
    Wedge, Daniel
    Holden, Eun-Jung
    Kovesi, Peter
    Clarke, Nick
    Vann, John
    MATHEMATICAL GEOSCIENCES, 2015, 47 (05) : 521 - 545
  • [32] Classification of Gold-Bearing Particles Using Visual Cues and Cost-Sensitive Machine Learning
    Tom Horrocks
    Daniel Wedge
    Eun-Jung Holden
    Peter Kovesi
    Nick Clarke
    John Vann
    Mathematical Geosciences, 2015, 47 : 521 - 545
  • [33] A genetic algorithm-based approach to cost-sensitive bankruptcy prediction
    Chen, Ning
    Ribeiro, Bernardete
    Vieira, Armando S.
    Duarte, Joao
    Neves, Joao C.
    EXPERT SYSTEMS WITH APPLICATIONS, 2011, 38 (10) : 12939 - 12945
  • [34] Imbalanced classification of mental workload using a cost-sensitive majority weighted minority oversampling strategy
    Jianhua Zhang
    Xiqing Cui
    Jianrong Li
    Rubin Wang
    Cognition, Technology & Work, 2017, 19 : 633 - 653
  • [35] Two-Stage Cost-Sensitive Learning for Software Defect Prediction
    Liu, Mingxia
    Miao, Linsong
    Zhang, Daoqiang
    IEEE TRANSACTIONS ON RELIABILITY, 2014, 63 (02) : 676 - 686
  • [36] Telugu Lyrics Based Classification By Using Naive Bayes
    Aziz, Abdul Md
    Ravula, Ganga Raju
    Shaik, Mobeen Taj
    Potharaju, Sravanthi
    2020 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND SIGNAL PROCESSING (AISP), 2020,
  • [37] Modifying Naive Bayes Classifier for Multinomial Text Classification
    Sharma, Neha
    Singh, Manoj
    2016 INTERNATIONAL CONFERENCE ON RECENT ADVANCES AND INNOVATIONS IN ENGINEERING (ICRAIE), 2016,
  • [38] Classification optimization for training a large dataset with Naive Bayes
    Thi Thanh Sang Nguyen
    Pham Minh Thu Do
    JOURNAL OF COMBINATORIAL OPTIMIZATION, 2020, 40 (01) : 141 - 169
  • [39] Imbalanced classification of mental workload using a cost-sensitive majority weighted minority oversampling strategy
    Zhang, Jianhua
    Cui, Xiqing
    Li, Jianrong
    Wang, Rubin
    COGNITION TECHNOLOGY & WORK, 2017, 19 (04) : 633 - 653
  • [40] Cost-sensitive learning for imbalanced medical data: a review
    Araf, Imane
    Idri, Ali
    Chairi, Ikram
    ARTIFICIAL INTELLIGENCE REVIEW, 2024, 57 (04)