Learning Label Specific Features for Multi-Label Classification

被引:168
作者
Huang, Jun [1 ,2 ,3 ]
Li, Guorong [1 ,2 ,3 ]
Huang, Qingming [1 ,2 ,3 ]
Wu, Xindong [4 ,5 ]
机构
[1] Univ Chinese Acad Sci, Sch Comp & Control Engn, Beijing 101480, Peoples R China
[2] Univ Chinese Acad Sci, Key Lab Big Data Min & Knowledge Management, Beijing, Peoples R China
[3] Chinese Acad Sci, Inst Comp Technol, Key Lab Intelligent Informat Proc, Beijing 100190, Peoples R China
[4] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei, Peoples R China
[5] Univ Vermont, Dept Comp Sci, Burlington, VT 05405 USA
来源
2015 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM) | 2015年
关键词
D O I
10.1109/ICDM.2015.67
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Binary relevance (BR) is a well-known framework for multi-label classification. It decomposes multi-label classification into binary (one-vs-rest) classification subproblems, one for each label. The BR approach is a simple and straightforward way for multi-label classification, but it still has several drawbacks. First, it does not consider label correlations. Second, each binary classifier may suffer from the issue of class-imbalance. Third, it can become computationally unaffordable for data sets with many labels. Several remedies have been proposed to solve these problems by exploiting label correlations between labels and performing label space dimension reduction. Meanwhile, inconsistency, another potential drawback of BR, is often ignored by researchers when they construct multi-label classification models. Inconsistency refers to the phenomenon that if an example belongs to more than one class label, then during the binary training stage, it can be considered as both positive and negative example simultaneously. This will mislead binary classifiers to learn suboptimal decision boundaries. In this paper, we seek to solve this problem by learning label specific features for each label. We assume that each label is only associated with a subset of features from the original feature set, and any two strongly correlated class labels can share more features with each other than two uncorrelated or weakly correlated ones. The proposed method can be applied as a feature selection method for multi-label learning and a general strategy to improve multi-label classification algorithms comprising a number of binary classifiers. Comparison with the state-of-the-art approaches manifests competitive performance of our proposed method.
引用
收藏
页码:181 / 190
页数:10
相关论文
共 45 条
[1]  
[Anonymous], 2009, Fast convex optimization algorithms for exact recovery of a corrupted low-rank matrix
[2]  
[Anonymous], 2010, P INT C MACH LEARN H
[3]  
[Anonymous], 2013, AAAI 13
[4]   Learning multi-label scene classification [J].
Boutell, MR ;
Luo, JB ;
Shen, XP ;
Brown, CM .
PATTERN RECOGNITION, 2004, 37 (09) :1757-1771
[5]   LIBSVM: A Library for Support Vector Machines [J].
Chang, Chih-Chung ;
Lin, Chih-Jen .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
[6]  
Chen Y.N., 2012, Advances in neural information processing systems, P1529
[7]   On label dependence and loss minimization in multi-label classification [J].
Dembczynski, Krzysztof ;
Waegeman, Willem ;
Cheng, Weiwei ;
Huellermeier, Eyke .
MACHINE LEARNING, 2012, 88 (1-2) :5-45
[8]   Leveraging Supervised Label Dependency Propagation for Multi-label Learning [J].
Fu, Bin ;
Xu, Guandong ;
Wang, Zhihai ;
Cao, Longbing .
2013 IEEE 13TH INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2013, :1061-1066
[9]   Multilabel classification via calibrated label ranking [J].
Fuernkranz, Johannes ;
Huellermeier, Eyke ;
Mencia, Eneldo Loza ;
Brinker, Klaus .
MACHINE LEARNING, 2008, 73 (02) :133-153
[10]  
Guo Y, 2011, IN C IND ENG ENG MAN, P1300, DOI 10.1109/IEEM.2011.6118126