Inter-class sparsity based discriminative least square regression

被引:191
作者
Wen, Jie [1 ,2 ]
Xu, Yong [1 ,2 ]
Li, Zuoyong [3 ]
Ma, Zhongli [4 ]
Xu, Yuanrong [1 ,2 ]
机构
[1] Harbin Inst Technol, Shenzhen Grad Sch, Biocomp Res Ctr, Shenzhen 518055, Guangdong, Peoples R China
[2] Harbin Inst Technol, Shenzhen Med Biometr Percept & Anal Engn Lab, Shenzhen 518055, Guangdong, Peoples R China
[3] Minjiang Univ, Fujian Prov Key Lab Informat Proc & Intelligent C, Fuzhou 350121, Fujian, Peoples R China
[4] Harbin Engn Univ, Coll Automat, Harbin 150001, Heilongjiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Least square regression; Inter-class sparsity; Multi-class classification; Supervised learning; FACE RECOGNITION; CLASSIFICATION; SELECTION;
D O I
10.1016/j.neunet.2018.02.002
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Least square regression is a very popular supervised classification method. However, two main issues greatly limit its performance. The first one is that it only focuses on fitting the input features to the corresponding output labels while ignoring the correlations among samples. The second one is that the used label matrix, i.e., zero-one label matrix is inappropriate for classification. To solve these problems and improve the performance, this paper presents a novel method, i.e., inter-class sparsity based discriminative least square regression (ICS_DLSR), for multi-class classification. Different from other methods, the proposed method pursues that the transformed samples have a common sparsity structure in each class. For this goal, an inter-class sparsity constraint is introduced to the least square regression model such that the margins of samples from the same class can be greatly reduced while those of samples from different classes can be enlarged. In addition, an error term with row-sparsity constraint is introduced to relax the strict zero-one label matrix, which allows the method to be more flexible in learning the discriminative transformation matrix. These factors encourage the method to learn a more compact and discriminative transformation for regression and thus has the potential to perform better than other methods. Extensive experimental results show that the proposed method achieves the best performance in comparison with other methods for multi-class classification. (c) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:36 / 47
页数:12
相关论文
共 55 条
[1]   Partial least squares regression and projection on latent structure regression (PLS Regression) [J].
Abdi, Herve .
WILEY INTERDISCIPLINARY REVIEWS-COMPUTATIONAL STATISTICS, 2010, 2 (01) :97-106
[2]  
[Anonymous], 2012, ARXIV12042358
[3]  
[Anonymous], 2011, P ADV NEUR INF PROC
[4]  
[Anonymous], 2016, Advances in Face Detection and Facial Image Analysis, DOI 10.1007/978-3-319-25958-1
[5]  
[Anonymous], 2012, MATH PROGRAM
[6]   Convex multi-task feature learning [J].
Argyriou, Andreas ;
Evgeniou, Theodoros ;
Pontil, Massimiliano .
MACHINE LEARNING, 2008, 73 (03) :243-272
[7]  
Bishop C.M., 2006, PATTERN RECOGN, V4, P738, DOI DOI 10.1117/1.2819119
[8]   OPTIMAL SELECTION OF REDUCED RANK ESTIMATORS OF HIGH-DIMENSIONAL MATRICES [J].
Bunea, Florentina ;
She, Yiyuan ;
Wegkamp, Marten H. .
ANNALS OF STATISTICS, 2011, 39 (02) :1282-1309
[9]  
Cai D, 2007, IEEE DATA MINING, P73, DOI 10.1109/ICDM.2007.89
[10]  
Cai X., 2013, P 23 INT JOINT C ART, P1240, DOI DOI 10.5555/2540128.2540307