Parameter optimization of kernel-based one-class classifier on imbalance learning

被引:0
作者
Zhuang, Ling [1 ]
Dai, Honghua [1 ]
机构
[1] School of Engineering and Information Technology, Deakin University, Burwood, VIC 3125
来源
Journal of Computers (Finland) | 2006年 / 1卷 / 07期
关键词
Imbalance learning; One-class classification framework; One-class support vector machine; Support vector data description(SVDD);
D O I
10.4304/jcp.1.7.32-40
中图分类号
学科分类号
摘要
Compared with conventional two-class learning schemes, one-class classification simply uses a single class in the classifier training phase. Applying one-class classification to learn from unbalanced data set is regarded as the recognition based learning and has shown to have the potential of achieving better performance. Similar to twoclass learning, parameter selection is a significant issue, especially when the classifier is sensitive to the parameters. For one-class learning scheme with the kernel function, such as one-class Support Vector Machine and Support Vector Data Description, besides the parameters involved in the kernel, there is another one-class specific parameter: the rejection rate v. In this paper, we proposed a general framework to involve the majority class in solving the parameter selection problem. In this framework, we first use the minority target class for training in the one-class classification stage; then we use both minority and majority class for estimating the generalization performance of the constructed classifier. This generalization performance is set as the optimization criteria. We employed the Grid search and Experiment Design search to attain various parameter settings. Experiments on UCI and Reuters text data show that the parameter optimized one-class classifiers outperform all the standard one-class learning schemes we examined. © 2006 ACADEMY PUBLISHER.
引用
收藏
页码:32 / 40
页数:8
相关论文
共 15 条
[1]  
Manevitz L.M., Yousef M., One-class svms for document classification, Journal of Machine Learning Research, 2, pp. 139-154, (2001)
[2]  
Raskutti B., Kowalczyk A., Extreme re-balancing for svms: A case study, SIGKDD Explorations, 6, pp. 60-69, (2004)
[3]  
Scholkopt B., Platt J.C., Shawe-Taylor J., Smola A.J., Williamson R.C., Estimating the support of a highdimensional distribution, Neural Computation, 13, pp. 1443-1471, (2001)
[4]  
David R.P.D., Tax M.J., Support vector domain description, Pattern Recognition Letters, 20, pp. 1191-1199, (1999)
[5]  
Tax D.M.J., One-class Classification, (2001)
[6]  
Tax D., Ddtools, the Data Description Toolbox For Matlab
[7]  
Japkowicz N., Supervised versus unsupervised binarylearning by feedforward neural networks, Machine Learning, 1-2, pp. 97-122, (2001)
[8]  
Zhang J., Bloedorn E., Rosen L., Venese D., Learning rules from highly unbalanced data sets, Proceedings of the IVth IEEE International Conference On Data Mining (ICDM'04), pp. 571-574, (2004)
[9]  
Tran Q.-A., Li X., Duan H., Efficient performance estimate for one-class support vector machine, Pattern Recognition Letters, 26, pp. 1174-1182, (2005)
[10]  
Tran Q.-A., Zhang Q., Li X., Evolving training model method for one-class svm, Proceeding of 2003 IEEE International Conference On Systems, (2003)