Expand globally, shrink locally: Discriminant multi-label learning with missing labels

被引:52
作者
Ma, Zhongchen [1 ,2 ]
Chen, Songcan [2 ,3 ]
机构
[1] Jiangsu Univ, Sch Comp Sci & Commun Engn, Zhenjiang 212013, Jiangsu, Peoples R China
[2] Nanjing Univ Aeronaut & Astronaut NUAA, Coll Comp Sci & Technol, Nanjing 211106, Peoples R China
[3] MIIT Key Lab Pattern Anal & Machine Intelligence, Nanjing, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-label learning; Missing labels; Local low-rank label structure; Global low-rank label structure; Label discriminant information; CLASSIFICATION;
D O I
10.1016/j.patcog.2020.107675
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In multi-label learning, the issue of missing labels brings a major challenge. Many methods attempt to recovery missing labels by exploiting low-rank structure of label matrix. However, these methods just utilize global low-rank label structure, ignore both local low-rank label structures and label discriminant information to some extent, leaving room for further performance improvement. In this paper, we develop a simple yet effective discriminant multi-label learning (DM2L) method for multi-label learning with missing labels. Specifically, we impose the low-rank structures on all the predictions of instances from the same labels (local shrinking of rank), and a maximally separated structure (high-rank structure) on the predictions of instances from different labels (global expanding of rank). In this way, these imposed low-rank structures can help modeling both local and global low-rank label structures, while the imposed high-rank structure can help providing more underlying discriminability. Our subsequent theoretical analysis also supports these intuitions. In addition, we provide a nonlinear extension via using kernel trick to enhance DM2L and establish a concave-convex objective to learn these models. Compared to the other methods, our method involves the fewest assumptions and only one hyper-parameter. Even so, extensive experiments show that our method still outperforms the state-of-the-art methods. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页数:10
相关论文
共 40 条
[1]  
[Anonymous], 2014, INT C MACH LEARN
[2]  
[Anonymous], Advances in neural information processing systems
[3]  
[Anonymous], 2011, NIPS
[4]   Large-Scale Machine Learning with Stochastic Gradient Descent [J].
Bottou, Leon .
COMPSTAT'2010: 19TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL STATISTICS, 2010, :177-186
[5]   Learning multi-label scene classification [J].
Boutell, MR ;
Luo, JB ;
Shen, XP ;
Brown, CM .
PATTERN RECOGNITION, 2004, 37 (09) :1757-1771
[6]  
Bucak SS, 2011, PROC CVPR IEEE
[7]  
Chen WF, 2013, IEEE INT CONF CON AU, P1274
[8]  
Clare A., 2001, EUR C PRINC DAT MIN, P42, DOI DOI 10.1007/3-540-44794-6_4
[9]  
Elisseeff A, 2002, ADV NEUR IN, V14, P681
[10]   Multilabel classification via calibrated label ranking [J].
Fuernkranz, Johannes ;
Huellermeier, Eyke ;
Mencia, Eneldo Loza ;
Brinker, Klaus .
MACHINE LEARNING, 2008, 73 (02) :133-153