Bagging Constraint Score for feature selection with pairwise constraints

被引:46
作者
Sun, Dan [1 ]
Zhang, Daoqiang [1 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Dept Comp Sci & Engn, Nanjing 210016, Peoples R China
基金
美国国家科学基金会;
关键词
Feature selection; Constraint Score; Pairwise constraints; Bagging; Ensemble learning; IMAGE RETRIEVAL; RELEVANCE; CLASSIFICATION; PREDICTION; FRAMEWORK; SVM;
D O I
10.1016/j.patcog.2009.12.011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Constraint Score is a recently proposed method for feature selection by using pairwise constraints which specify whether a pair of instances belongs to the same class or not. It has been shown that the Constraint Score, with only a small amount of pairwise constraints, achieves comparable performance to those fully supervised feature selection methods such as Fisher Score. However, one major disadvantage of the Constraint Score is that its performance is dependent on a good selection on the composition and cardinality of constraint set, which is very challenging in practice. In this work, we address the problem by importing Bagging into Constraint Score and a new method called Bagging Constraint Score (BCS) is proposed. Instead of seeking one appropriate constraint set for single Constraint Score, in BCS we perform multiple Constraint Score, each of which uses a bootstrapped subset of original given constraint set. Diversity analysis on individuals of ensemble shows that resampling pairwise constraints is helpful for simultaneously improving accuracy and diversity of individuals. We conduct extensive experiments on a series of high-dimensional datasets from UCI repository and gene databases, and the experimental results validate the effectiveness of the proposed method. (C) 2009 Elsevier Ltd. All rights reserved.
引用
收藏
页码:2106 / 2118
页数:13
相关论文
共 51 条
  • [21] Ho TK, 1998, IEEE T PATTERN ANAL, V20, P832, DOI 10.1109/34.709601
  • [22] Performance of feature-selection methods in the classification of high-dimension data
    Hua, Jianping
    Tembe, Waibhav D.
    Dougherty, Edward R.
    [J]. PATTERN RECOGNITION, 2009, 42 (03) : 409 - 424
  • [23] Feature selection: Evaluation, application, and small sample performance
    Jain, A
    Zongker, D
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1997, 19 (02) : 153 - 158
  • [24] KIRA K, 1992, AAAI-92 PROCEEDINGS : TENTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, P129
  • [25] Wrappers for feature subset selection
    Kohavi, R
    John, GH
    [J]. ARTIFICIAL INTELLIGENCE, 1997, 97 (1-2) : 273 - 324
  • [26] Feature selection based on loss-margin of nearest neighbor classification
    Li, Yun
    Lu, Bao-Liang
    [J]. PATTERN RECOGNITION, 2009, 42 (09) : 1914 - 1921
  • [27] Feature selection with dynamic mutual information
    Liu, Huawen
    Sun, Jigui
    Liu, Lei
    Zhang, Huijie
    [J]. PATTERN RECOGNITION, 2009, 42 (07) : 1330 - 1339
  • [28] Ensemble component selection for improving ICA based microarray data prediction models
    Liu, Kun-Hong
    Li, Bo
    Zhang, Jun
    Du, Ji-Xiang
    [J]. PATTERN RECOGNITION, 2009, 42 (07) : 1274 - 1283
  • [29] Melville P., 2004, P 21 INT C MACHINE L, P74
  • [30] An improvement on floating search algorithms for feature subset selection
    Nakariyakul, Songyot
    Casasent, David P.
    [J]. PATTERN RECOGNITION, 2009, 42 (09) : 1932 - 1940