Learning from labeled and unlabeled data using a minimal number of queries

被引:24
作者
Kothari, R [1 ]
Jain, V [1 ]
机构
[1] Indian Inst Technol, IBM India Res Lab, New Delhi 110016, India
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2003年 / 14卷 / 06期
关键词
expectation-maximization; genetic algorithms (GAs); learning; supervised learning; unsupervised learning;
D O I
10.1109/TNN.2003.820446
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The considerable time and expense required for labeling data has prompted the development of algorithms which maximize the classification accuracy for a given amount of labeling effort. On the one hand, the effort has been to develop the so-called active learning" algorithms which sequentially choose the pat_ terns to be explictly labeled so as to realize the maximum infor mation gain from each labeling. On the other hand, the effort has been to develop algorithms that can learn from labeled as well as the more abundant unlabeled data. Proposed in this paper is an algorithm that integrates the benefits of active learning with the benefits of learning from labeled and unlabeled data. Our approach is based on reversing the roles of the labeled and unlabeled data. Specifically, we use a Genetic Algorithm (GA) to iteratively refine the class membership of the unlabeled patterns so that the maximum a posteriori (MAP) based predicted labels of the patterns in the labeled dataset are in agreement with the known labels. This reversal of the role of labeled and unlabeled patterns leads to an implicit class assignment of the unlabeled patterns. For active learning, we use a subset of the GA population to construct multiple MAP classifiers. Points in the input space where there is maximal disagreement amongst these classifiers are then selected for explicit labeling. The learning from labeled and unlabeled data and active learning phases are interlaced and together provide accurate classification while minimizing the labeling effort.
引用
收藏
页码:1496 / 1505
页数:10
相关论文
共 38 条
  • [1] Asoh H, 1994, LECT NOTES COMPUT SC, V866, P88
  • [2] Bishop C. M., 1995, NEURAL NETWORKS PATT
  • [3] Blake C.L., 1998, UCI repository of machine learning databases
  • [4] Blum A., 1998, Proceedings of the Eleventh Annual Conference on Computational Learning Theory, P92, DOI 10.1145/279943.279962
  • [5] ON THE EXPONENTIAL VALUE OF LABELED SAMPLES
    CASTELLI, V
    COVER, TM
    [J]. PATTERN RECOGNITION LETTERS, 1995, 16 (01) : 105 - 111
  • [6] CERONI A, 2001, CONVERGENCE TIME MOD
  • [7] Cohn D. A., 1995, Advances in Neural Information Processing Systems 7, P705
  • [8] CORTES C, 1995, MACH LEARN, V20, P273, DOI 10.1023/A:1022627411411
  • [9] Fedorov VV., 1972, THEORY OPTIMAL EXPT
  • [10] Ganesalingam S., 1979, Journal of Statistical Computation and Simulation, V9, P151, DOI 10.1080/00949657908810306