A bagging SVM to learn from positive and unlabeled examples

被引:201
作者
Mordelet, F. [1 ]
Vert, J. -P. [2 ,3 ,4 ]
机构
[1] Duke Univ, Durham, NC 27708 USA
[2] Mines ParisTech, Ctr Computat Biol, F-77300 Fontainebleau, France
[3] Inst Curie, F-75005 Paris, France
[4] INSERM, U900, F-75005 Paris, France
关键词
PU learning; Bagging; SVM; CLASSIFICATION;
D O I
10.1016/j.patrec.2013.06.010
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the problem of learning a binary classifier from a training set of positive and unlabeled examples, both in the inductive and in the transductive setting. This problem, often referred to as PU learning, differs from the standard supervised classification problem by the lack of negative examples in the training set. It corresponds to an ubiquitous situation in many applications such as information retrieval or gene ranking, when we have identified a set of data of interest sharing a particular property, and we wish to automatically retrieve additional data sharing the same property among a large and easily available pool of unlabeled data. We propose a new method for PU learning with a conceptually simple implementation based on bootstrap aggregating (bagging) techniques: the algorithm iteratively trains many binary classifiers to discriminate the known positive examples from random subsamples of the unlabeled set, and averages their predictions. We show theoretically and experimentally that the method can match and even outperform the performance of state-of-the-art methods for PU learning, particularly when the number of positive examples is limited and the fraction of negatives among the unlabeled examples is small. The proposed method can also run considerably faster than state-of-the-art methods, particularly when the set of unlabeled examples is large. (C) 2013 Elsevier B.V. All rights reserved.
引用
收藏
页码:201 / 209
页数:9
相关论文
共 31 条
[21]  
Pelckmans K., 2009, ESANN 2009
[22]   RegulonDB (version 5.0):: Escherichia coli K-12 transcriptional regulatory network, operon organization, and growth conditions [J].
Salgado, Heladia ;
Gama-Castro, Socorro ;
Peralta-Gil, Martin ;
Diaz-Peredo, Edgar ;
Sanchez-Solano, Fabiola ;
Santos-Zavaleta, Alberto ;
Martinez-Flores, Irma ;
Jimenez-Jacinto, Veronica ;
Bonavides-Martinez, Cesar ;
Segura-Salazar, Juan ;
Martinez-Antonio, Agustino ;
Collado-Vides, Julio .
NUCLEIC ACIDS RESEARCH, 2006, 34 :D394-D397
[23]   Estimating the support of a high-dimensional distribution [J].
Schölkopf, B ;
Platt, JC ;
Shawe-Taylor, J ;
Smola, AJ ;
Williamson, RC .
NEURAL COMPUTATION, 2001, 13 (07) :1443-1471
[24]   A Neyman-Pearson approach to statistical learning [J].
Scott, C ;
Nowak, R .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2005, 51 (11) :3806-3819
[25]   SVM-HUSTLE - an iterative semi-supervised machine learning approach for pairwise protein remote homology detection [J].
Shah, Anuj R. ;
Oehmen, Christopher S. ;
Webb-Robertson, Bobbie-Jo .
BIOINFORMATICS, 2008, 24 (06) :783-790
[26]  
Sriphaew K, 2009, LECT NOTES ARTIF INT, V5476, P62, DOI 10.1007/978-3-642-01307-2_9
[27]   Sparseness of support vector machines [J].
Steinwart, I .
JOURNAL OF MACHINE LEARNING RESEARCH, 2004, 4 (06) :1071-1105
[28]  
Vert R, 2006, J MACH LEARN RES, V7, P817
[29]  
Welling M, 2009, P 12 INT C ART INT S, P464
[30]   PEBL: Web page classification without negative examples [J].
Yu, HJ ;
Han, JW ;
Chang, KCC .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2004, 16 (01) :70-81