Kernels for Generalized Multiple-Instance Learning

被引:11
|
作者
Tao, Qingping [1 ]
Scott, Stephen D. [2 ]
Vinodchandran, N. V. [2 ]
Osugi, Thomas Takeo [3 ]
Mueller, Brandon [4 ]
机构
[1] GC Image LLC, Lincoln, NE 68505 USA
[2] Univ Nebraska, Dept Comp Sci, Lincoln, NE 68588 USA
[3] Sphere Commun, Lincolnshire, IL 60069 USA
[4] Gallup Inc, Omaha, NE 68102 USA
基金
美国国家科学基金会; 美国国家卫生研究院;
关键词
Kernels; support vector machines; generalized multiple-instance learning; content-based image retrieval; biological sequence analysis; fully polynomial randomized approximation schemes;
D O I
10.1109/TPAMI.2007.70846
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The multiple-instance learning (MIL) model has been successful in numerous application areas. Recently, a generalization of this model and an algorithm for it have been introduced, showing significant advantages over the conventional MIL model on certain application areas. Unfortunately, that algorithm is not scalable to high dimensions. We adapt that algorithm to one that uses a support vector machine with our new kernel k(A). This reduces the time complexity from exponential in the dimension to polynomial. Computing our new kernel is equivalent to counting the number of boxes in a discrete bounded space that contain at least one point from each of two multisets. We first show that this problem is #P-complete and then present a fully polynomial randomized approximation scheme (FPRAS) for it. We then extend k(A) by enriching its representation into a new kernel k(min) and also consider a normalized version of k(A) that we call k(A/V) (which may or may not be a kernel but whose approximation yielded positive semidefinite Gram matrices in practice). We then empirically evaluate all three measures on data from content-based image retrieval, biological sequence analysis, and the Musk data sets. We found that our kernels performed well on all data sets relative to algorithms in the conventional MIL model.
引用
收藏
页码:2084 / 2097
页数:14
相关论文
共 50 条
  • [1] ON GENERALIZED MULTIPLE-INSTANCE LEARNING
    Scott, Stephen
    Zhang, Jun
    Brown, Joshua
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE AND APPLICATIONS, 2005, 5 (01) : 21 - 35
  • [2] An extended kernel for generalized multiple-instance learning
    Tao, QP
    Scott, S
    Vinodchandran, NV
    Osugi, TT
    Mueller, B
    ICTAI 2004: 16TH IEEE INTERNATIONALCONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2004, : 272 - 277
  • [3] Detecting Multiple Myeloma via Generalized Multiple-Instance Learning
    Hering, Jan
    Kybic, Jan
    Lambert, Lukas
    MEDICAL IMAGING 2018: IMAGE PROCESSING, 2018, 10574
  • [4] Compact Multiple-Instance Learning
    Chai, Jing
    Liu, Weiwei
    Tsang, Ivor W.
    Shen, Xiaobo
    CIKM'17: PROCEEDINGS OF THE 2017 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2017, : 2007 - 2010
  • [5] On multiple-instance learning of halfspaces
    Diochnos, D. I.
    Sloan, R. H.
    Turan, Gy
    INFORMATION PROCESSING LETTERS, 2012, 112 (23) : 933 - 936
  • [6] A framework for multiple-instance learning
    Maron, O
    Lozano-Perez, T
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 10, 1998, 10 : 570 - 576
  • [7] UNSUPERVISED MULTIPLE-INSTANCE LEARNING FOR INSTANCE SEARCH
    Wang, Zhenzhen
    Yuan, Junsong
    2018 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2018,
  • [8] MULTIPLE-INSTANCE LEARNING WITH PAIRWISE INSTANCE SIMILARITY
    Yuan, Liming
    Liu, Jiafeng
    Tang, Xianglong
    INTERNATIONAL JOURNAL OF APPLIED MATHEMATICS AND COMPUTER SCIENCE, 2014, 24 (03) : 567 - 577
  • [9] Salient Instance Selection for Multiple-Instance Learning
    Yuan, Liming
    Liu, Songbo
    Huang, Qingcheng
    Liu, Jiafeng
    Tang, Xianglong
    NEURAL INFORMATION PROCESSING, ICONIP 2012, PT III, 2012, 7665 : 58 - 67
  • [10] Multiple-Instance Learning from Distributions
    Doran, Gary
    Ray, Soumya
    JOURNAL OF MACHINE LEARNING RESEARCH, 2016, 17