Chunk incremental learning for cost-sensitive hinge loss support vector machine

被引:41
作者
Gu, Bin [1 ,2 ,3 ]
Quan, Xin [3 ]
Gu, Yunhua [3 ]
Sheng, Victor S. [4 ]
Zheng, Guansheng [3 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Jiangsu Key Lab Big Data Anal Technol B DAT, Nanjing, Jiangsu, Peoples R China
[2] Jiangsu Collaborat Innovat Ctr Atmospher Environm, Nanjing, Jiangsu, Peoples R China
[3] Nanjing Univ Informat Sci & Technol, Sch Comp & Software, Nanjing, Jiangsu, Peoples R China
[4] Univ Cent Arkansas, Dept Comp Sci, Conway, AR USA
基金
美国国家科学基金会; 中国国家自然科学基金;
关键词
Cost-sensitive learning; Chunk incremental learning; Hinge loss; Support vector machines; ONLINE; CLASSIFIERS; ALGORITHM;
D O I
10.1016/j.patcog.2018.05.023
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cost-sensitive learning can be found in many real-world applications and represents an important learning paradigm in machine learning. The recently proposed cost-sensitive hinge loss support vector machine (CSHL-SVM) guarantees consistency with the cost-sensitive Bayes risk, and this technique provides better generalization accuracy compared to traditional cost-sensitive support vector machines. In practice, data typically appear in the form of sequential chunks, also called an on-line scenario. However, conventional batch learning algorithms waste a considerable amount of time under the on-line scenario due to re-training of a model from scratch. To make CSHL-SVM more practical for the on-line scenario, we propose a chunk incremental learning algorithm for CSHL-SVM, which can update a trained model without re-training from scratch when incorporating a chunk of new samples. Our method is efficient because it can update the trained model for not only one sample at a time but also multiple samples at a time. Our experimental results on a variety of datasets not only confirm the effectiveness of CSHL-SVM but also show that our method is more efficient than the batch algorithm of CSHL-SVM and the incremental learning method of CSHL-SVM only for a single sample. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:196 / 208
页数:13
相关论文
共 48 条
[1]  
Alcalá-Fdez J, 2011, J MULT-VALUED LOG S, V17, P255
[2]  
[Anonymous], 2014, Advances in Neural Information Processing Systems 27
[3]  
[Anonymous], 2012, MATRIX COMPUTATIONS
[4]  
[Anonymous], 2013, ICML
[5]  
[Anonymous], 2002, Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond
[6]  
[Anonymous], INT J ENG SCI
[7]  
[Anonymous], INT JOINT C ART INT
[8]  
[Anonymous], THESIS
[9]  
[Anonymous], 2011, ACM T INTEL SYST TEC, DOI DOI 10.1145/1961189.1961199
[10]  
Bordes A, 2005, J MACH LEARN RES, V6, P1579