Efficient Decomposition Selection for Multi-class Classification

被引:1
作者
Chen, Yawen [1 ]
Wen, Zeyi [2 ]
He, Bingsheng [3 ,4 ]
Chen, Jian [1 ]
机构
[1] South China Univ Technol, Guangzhou 510000, Guangdong, Peoples R China
[2] Univ Western Australia, Crawley, WA 6009, Australia
[3] Natl Univ Singapore, Sch Comp, Singapore 119077, Singapore
[4] NUS Ctr Trust Internet & Community, Singapore 119077, Singapore
基金
中国国家自然科学基金;
关键词
Indexes; Matrix decomposition; Kernel; Codes; Training; Support vector machines; Probability distribution; Machine learning; multi-class classification; decomposition method; MATRIX;
D O I
10.1109/TKDE.2021.3130239
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Choosing a decomposition method for multi-class classification is an important trade-off between efficiency and predictive accuracy. Trying all the decomposition methods to find the best one is too time-consuming for many applications, while choosing the wrong one may result in large loss on predictive accuracy. In this paper, we propose an automatic decomposition method selection approach called "D-Chooser", which is lightweight and can choose the best decomposition method accurately. D-Chooser is equipped with our proposed difficulty index which consists of sub-metrics including distribution divergence, overlapping regions, unevenness degree and relative size of the solution space. The difficulty index has two intriguing properties: 1) fast to compute and 2) measuring multi-class problems comprehensively. Extensive experiments on real-world multi-class problems show that D-Chooser achieves an accuracy of 80.56% in choosing the best decomposition method. It can choose the best method in just a few seconds, while existing approaches verify the effectiveness of a decomposition method often takes a few hours. We also provide case studies on Kaggle competitions and the results confirm that D-Chooser is able to choose a better decomposition method than the winning solutions.
引用
收藏
页码:3751 / 3764
页数:14
相关论文
共 40 条
[1]  
Aly M., 2005, Neural Netw., V19, P1
[2]  
Anguita D., 2012, P 4 INT WORKSH AMB A, V7657, P216
[3]  
[Anonymous], 2012, NeurIPS
[4]   MINIMUM HELLINGER DISTANCE ESTIMATES FOR PARAMETRIC MODELS [J].
BERAN, R .
ANNALS OF STATISTICS, 1977, 5 (03) :445-463
[5]   Birdsnap: Large-scale Fine-grained Visual Categorization of Birds [J].
Berg, Thomas ;
Liu, Jiongxin ;
Lee, Seung Woo ;
Alexander, Michelle L. ;
Jacobs, David W. ;
Belhumeur, Peter N. .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :2019-2026
[6]  
Chen YW, 2018, AAAI CONF ARTIF INTE, P8061
[7]  
Cotter A., 2011, P 17 ACM SIGKDD INT, P805, DOI [10.1145/2020408.2020548, DOI 10.1145/2020408.2020548]
[8]  
Dietterich TG, 1994, J ARTIF INTELL RES, V2, P263
[9]  
DON FJH, 1987, LINEAR ALGEBRA APPL, V93, P1
[10]  
Falkner S., 2017, NIPS 2017 Bayesian optimization workshop, P4