Feature selection for least squares projection twin support vector machine

被引:26
作者
Guo, Jianhui [1 ,5 ]
Yi, Ping [2 ]
Wang, Ruili [3 ]
Ye, Qiaolin [4 ]
Zhao, Chunxia [1 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing, Jiangsu, Peoples R China
[2] Southeast Univ, Sch Instrument Sci & Engn, Nanjing, Jiangsu, Peoples R China
[3] Massey Univ, Sch Engn & Adv Technol, Auckland, New Zealand
[4] Nanjing Forestry Univ, Sch Informat Technol, Nanjing, Jiangsu, Peoples R China
[5] Nanjing Univ Sci & Technol, Jiangsu Key Lab Image & Video Understanding Socia, Nanjing, Jiangsu, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Twin Support Vector Machine; Least Squares Projection Twin Support; Vector Machine; Feature selection; SVM;
D O I
10.1016/j.neucom.2014.05.040
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a new feature selection approach for the recently proposed Least Squares Projection Twin Support Vector Machine (LSPTSVM) for binary classification. 1-norm is used in our feature selection objective so that only non-zero elements in weight vectors will be chosen as selected features. Also, the Tikhonov regularization term is incorporated to the objective of our approach to reduce the singularity problems of Quadratic Programming Problems (QPPs), and then to minimize its 1-norm measure. This approach leads to a strong feature suppression capability, called as Feature Selection for Least Squares Projection Twin Support Vector Machine (FLSPTSVM). The solutions of FLSPTSVM can be obtained by solving two smaller QPPS arising from two primal QPPs as opposed to two dual ones in Twin Support Vector Machine (TWSVM). Thus, FLSPTSVM is capable of generating sparse solutions. This means that FLSPTSVM can reduce the number of input features for a linear case. Our linear FLSPTSVM can also be extended to a nonlinear case with the kernel trick. When a nonlinear classifier is used, the number of kernel functions required for the classifier is reduced. Our experiments on publicly available datasets demonstrate that our FLSPTSVM has comparable classification accuracy to that of LSPTSVM and obtains sparse solutions. (C) 2014 Elsevier B.V. All rights reserved.
引用
收藏
页码:174 / 183
页数:10
相关论文
共 36 条
[1]  
[Anonymous], P 11 INT C ART INT S
[2]  
Blake C. L., 1998, Uci repository of machine learning databases
[3]   A tutorial on Support Vector Machines for pattern recognition [J].
Burges, CJC .
DATA MINING AND KNOWLEDGE DISCOVERY, 1998, 2 (02) :121-167
[4]   Support vector machines experts for time series forecasting [J].
Cao, LJ .
NEUROCOMPUTING, 2003, 51 :321-339
[5]   Regularized least squares fisher linear discriminant with applications to image recognition [J].
Chen, Xiaobo ;
Yang, Jian ;
Mao, Qirong ;
Han, Fei .
NEUROCOMPUTING, 2013, 122 :521-534
[6]   Recursive robust least squares support vector regression based on maximum correntropy criterion [J].
Chen, Xiaobo ;
Yang, Jian ;
Liang, Jun ;
Ye, Qiaolin .
NEUROCOMPUTING, 2012, 97 :63-73
[7]   Recursive projection twin support vector machine via within-class variance minimization [J].
Chen, Xiaobo ;
Yang, Jian ;
Ye, Qiaolin ;
Liang, Jun .
PATTERN RECOGNITION, 2011, 44 (10-11) :2643-2655
[8]  
Christianini N., 2000, INTRO SUPPORT VECTOR, P189
[9]  
Ding S., 2012, ARTIF INTELL REV
[10]   Graph-Based Learning via Auto-Grouped Sparse Regularization and Kernelized Extension [J].
Fang, Yuqiang ;
Wang, Ruili ;
Dai, Bin ;
Wu, Xindong .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2015, 27 (01) :142-154