Active Property Testing

被引:23
作者
Balcan, Maria-Florina [1 ]
Blais, Eric [2 ]
Blum, Avrim [2 ]
Yang, Liu [2 ]
机构
[1] Georgia Inst Technol, Sch Comp Sci, Atlanta, GA 30332 USA
[2] Carnegie Mellon Univ Pittsburgh, Sch Comp Sci, Pittsburgh, PA USA
来源
2012 IEEE 53RD ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS) | 2012年
基金
美国国家科学基金会;
关键词
Property testing; Active learning; Boolean functions; Linear threshold functions; Unions of intervals; LEARNABILITY;
D O I
10.1109/FOCS.2012.64
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
One motivation for property testing of boolean functions is the idea that testing can provide a fast preprocessing step before learning. However, in most machine learning applications, it is not possible to request for labels of arbitrary examples constructed by an algorithm. Instead, the dominant query paradigm in applied machine learning, called active learning, is one where the algorithm may query for labels, but only on points in a given (polynomial-sized) unlabeled sample, drawn from some underlying distribution D. In this work, we bring this well-studied model to the domain of testing. We develop both general results for this active testing model as well as efficient testing algorithms for several important properties for learning, demonstrating that testing can still yield substantial benefits in this restricted setting. For example, we show that testing unions of d intervals can be done with O(1) label requests in our setting, whereas it is known to require Omega(d) labeled examples for learning (and Omega(root d) for passive testing [22] where the algorithm must pay for every example drawn from D). In fact, our results for testing unions of intervals also yield improvements on prior work in both the classic query model (where any point in the domain can be queried) and the passive testing model as well. For the problem of testing linear separators in R-n over the Gaussian distribution, we show that both active and passive testing can be done with O(root n) queries, substantially less than the Omega(n) needed for learning, with near-matching lower bounds. We also present a general combination result in this model for building testable properties out of others, which we then use to provide testers for a number of assumptions used in semi-supervised learning. In addition to the above results, we also develop a general notion of the testing dimension of a given property with respect to a given distribution, that we show characterizes (up to constant factors) the intrinsic number of label requests needed to test that property. We develop such notions for both the active and passive testing models. We then use these dimensions to prove a number of lower bounds, including for linear separators and the class of dictator functions.
引用
收藏
页码:21 / 30
页数:10
相关论文
共 31 条
  • [1] [Anonymous], 2006, BOOK REV IEEE T NEUR
  • [2] A BERNSTEIN-TYPE INEQUALITY FOR U-STATISTICS AND U-PROCESSES
    ARCONES, MA
    [J]. STATISTICS & PROBABILITY LETTERS, 1995, 22 (03) : 239 - 247
  • [3] Balcan M.-F., 2012, CORR
  • [4] Balcan Maria-Florina, 2006, P 23 ICML
  • [5] Bartlett P. L., 2003, Journal of Machine Learning Research, V3, P463, DOI 10.1162/153244303321897690
  • [6] BAUM E, 1993, P IEEE INT JOINT C N
  • [7] Beygelzimer Alina, 2009, P 26 ICML
  • [8] Blais E, 2009, ACM S THEORY COMPUT, P151
  • [9] Blum A., 1994, Proceedings of the Twenty-Sixth Annual ACM Symposium on the Theory of Computing, P253, DOI 10.1145/195058.195147
  • [10] LEARNABILITY AND THE VAPNIK-CHERVONENKIS DIMENSION
    BLUMER, A
    EHRENFEUCHT, A
    HAUSSLER, D
    WARMUTH, MK
    [J]. JOURNAL OF THE ACM, 1989, 36 (04) : 929 - 965