FEATURE SELECTION WITH WEIGHTED CONDITIONAL MUTUAL INFORMATION

被引:0
作者
Celik, Ceyhun [1 ]
Bilge, Hasan Sakir [1 ]
机构
[1] Gazi Univ, Muhendislik Fak, Bilgisayar Muhendisligi Bolumu, Ankara, Turkey
来源
JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY | 2015年 / 30卷 / 04期
关键词
Feature selection; conditional mutual information; maximum dependency; minimum redundancy; INPUT FEATURE-SELECTION; CLASSIFICATION; RELEVANCE;
D O I
暂无
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Huge data processing and extracting the meaningful information from those data is one of the important topics in data mining. In practice, it cannot be known whether present data are relevant to the problem, and irrelevant data increase the complexity of the prospective model. Dimensionality reduction approaches are applied to the problem parameters to build simpler and low cost models. Information theory based mutual information approaches are commonly used on dimensionality reduction. In these approaches, it is aimed to have the minimum redundancy and maximum dependency in outputs between inputs in the subset obtained from the data set in the execution of the size degradation. However the heuristic functions which are used in the proposed approaches to ensure this condition, control the relation between dependency and redundancy with the fixed parameter and problem-independent. In this study, a new mutual information approach is proposed. The heuristic function used in this proposed approach weights the effectiveness of redundancy on selection by evaluating the relationship between mutual information of features with class and mutual information of features among themselves. Similarly, both conditional mutual information and mutual information are calculated for maximum dependency. Thus the proposed heuristic function presents a dynamic approach to variety of problems. The obtained results of the tests point out the success of the proposed approach.
引用
收藏
页码:585 / 596
页数:12
相关论文
共 30 条
[1]   A Novel Weighted Support Vector Machine Based on Particle Swarm Optimization for Gene Selection and Tumor Classification [J].
Abdi, Mohammad Javad ;
Hosseini, Seyed Mohammad ;
Rezghi, Mansoor .
COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE, 2012, 2012
[2]   Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays [J].
Alon, U ;
Barkai, N ;
Notterman, DA ;
Gish, K ;
Ybarra, S ;
Mack, D ;
Levine, AJ .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 1999, 96 (12) :6745-6750
[3]  
Alpaydin E, 2021, Introduction to machine learning
[4]  
Bache K, 2013, UCI machine learning repository
[5]   USING MUTUAL INFORMATION FOR SELECTING FEATURES IN SUPERVISED NEURAL-NET LEARNING [J].
BATTITI, R .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (04) :537-550
[6]   Eye Movement Analysis for Activity Recognition Using Electrooculography [J].
Bulling, Andreas ;
Ward, Jamie A. ;
Gellersen, Hans ;
Troester, Gerhard .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2011, 33 (04) :741-753
[7]   Prediction of lysine ubiquitination with mRMR feature selection and analysis [J].
Cai, Yudong ;
Huang, Tao ;
Hu, Lele ;
Shi, Xiaohe ;
Xie, Lu ;
Li, Yixue .
AMINO ACIDS, 2012, 42 (04) :1387-1395
[8]   A survey on feature selection methods [J].
Chandrashekar, Girish ;
Sahin, Ferat .
COMPUTERS & ELECTRICAL ENGINEERING, 2014, 40 (01) :16-28
[9]   Conditional Mutual Information-Based Feature Selection Analyzing for Synergy and Redundancy [J].
Cheng, Hongrong ;
Qin, Zhiguang ;
Feng, Chaosheng ;
Wang, Yong ;
Li, Fagen .
ETRI JOURNAL, 2011, 33 (02) :210-218
[10]   Estimating optimal feature subsets using efficient estimation of high-dimensional mutual information [J].
Chow, TWS ;
Huang, D .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2005, 16 (01) :213-224