Kernels for Generalized Multiple-Instance Learning

被引:11
|
作者
Tao, Qingping [1 ]
Scott, Stephen D. [2 ]
Vinodchandran, N. V. [2 ]
Osugi, Thomas Takeo [3 ]
Mueller, Brandon [4 ]
机构
[1] GC Image LLC, Lincoln, NE 68505 USA
[2] Univ Nebraska, Dept Comp Sci, Lincoln, NE 68588 USA
[3] Sphere Commun, Lincolnshire, IL 60069 USA
[4] Gallup Inc, Omaha, NE 68102 USA
基金
美国国家科学基金会; 美国国家卫生研究院;
关键词
Kernels; support vector machines; generalized multiple-instance learning; content-based image retrieval; biological sequence analysis; fully polynomial randomized approximation schemes;
D O I
10.1109/TPAMI.2007.70846
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The multiple-instance learning (MIL) model has been successful in numerous application areas. Recently, a generalization of this model and an algorithm for it have been introduced, showing significant advantages over the conventional MIL model on certain application areas. Unfortunately, that algorithm is not scalable to high dimensions. We adapt that algorithm to one that uses a support vector machine with our new kernel k(A). This reduces the time complexity from exponential in the dimension to polynomial. Computing our new kernel is equivalent to counting the number of boxes in a discrete bounded space that contain at least one point from each of two multisets. We first show that this problem is #P-complete and then present a fully polynomial randomized approximation scheme (FPRAS) for it. We then extend k(A) by enriching its representation into a new kernel k(min) and also consider a normalized version of k(A) that we call k(A/V) (which may or may not be a kernel but whose approximation yielded positive semidefinite Gram matrices in practice). We then empirically evaluate all three measures on data from content-based image retrieval, biological sequence analysis, and the Musk data sets. We found that our kernels performed well on all data sets relative to algorithms in the conventional MIL model.
引用
收藏
页码:2084 / 2097
页数:14
相关论文
共 50 条
  • [21] A Note on Learning from Multiple-Instance Examples
    Avrim Blum
    Adam Kalai
    Machine Learning, 1998, 30 : 23 - 29
  • [22] Multiple-instance learning as a classifier combining problem
    Li, Yan
    Tax, David M. J.
    Duin, Robert P. W.
    Loog, Marco
    PATTERN RECOGNITION, 2013, 46 (03) : 865 - 874
  • [23] A note on learning from multiple-instance examples
    Blum, A
    Kalai, A
    MACHINE LEARNING, 1998, 30 (01) : 23 - 29
  • [24] Multiple-Instance Active Learning for Image Categorization
    Liu, Dong
    Hua, Xian-Sheng
    Yang, Linjun
    Zhang, Hong-Jiang
    ADVANCES IN MULTIMEDIA MODELING, PROCEEDINGS, 2009, 5371 : 239 - +
  • [25] MILD: Multiple-Instance Learning via Disambiguation
    Li, Wu-Jun
    Yeung, Dit-Yan
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2010, 22 (01) : 76 - 89
  • [26] Multiple-instance learning via random walk
    Wang, Dong
    Li, Jianmin
    Zhang, Bo
    MACHINE LEARNING: ECML 2006, PROCEEDINGS, 2006, 4212 : 473 - 484
  • [27] MIForests: Multiple-Instance Learning with Randomized Trees
    Leistner, Christian
    Saffari, Amir
    Bischof, Horst
    COMPUTER VISION - ECCV 2010, PT VI, 2010, 6316 : 29 - 42
  • [28] Fast Bundle Algorithm for Multiple-Instance Learning
    Bergeron, Charles
    Moore, Gregory
    Zaretzki, Jed
    Breneman, Curt M.
    Bennett, Kristin P.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (06) : 1068 - 1079
  • [29] Multiple-Instance Learning with Instance Selection via Constructive Covering Algorithm
    Zhang, Yanping
    Zhang, Heng
    Wei, Huazhen
    Tang, Jie
    Zhao, Shu
    TSINGHUA SCIENCE AND TECHNOLOGY, 2014, 19 (03) : 285 - 292
  • [30] Learning Instance Concepts from Multiple-Instance Data with Bags as Distributions
    Doran, Gary
    Ray, Soumya
    PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 1802 - 1808