SALIC: Social Active Learning for Image Classification

被引:4
作者
Chatzilari, Elisavet [1 ]
Nikolopoulos, Spiros [1 ]
Kompatsiaris, Yiannis [1 ]
Kittler, Josef [2 ]
机构
[1] Ctr Res & Technol Hellas, Inst Informat Technol, Thessaloniki 57001, Greece
[2] Univ Surrey, Ctr Vis Speech & Signal Proc, Guildford GU2 7XH, Surrey, England
关键词
Active learning; image classification; large scale; multi-modal fusion; social context; user tagged images;
D O I
10.1109/TMM.2016.2565440
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we present SALIC, an active learning method for selecting the most appropriate user tagged images to expand the training set of a binary classifier. The process of active learning can be fully automated in this social context by replacing the human oraclewith the images' tags. However, their noisy nature adds further complexity to the sample selection process since, apart from the images' informativeness (i.e., how much they are expected to inform the classifier if we knew their label), our confidence about their actual label should also be maximized (i.e., how certain the oracle is on the images' true contents). The main contribution of this work is in proposing a probabilistic approach for jointly maximizing the two aforementioned quantities. In the examined noisy context, the oracle's confidence is necessary to provide a contextual-based indication of the images' true contents, while the samples' informativeness is required to reduce the computational complexity and minimize the mistakes of the unreliable oracle. To prove this, first, we show that SALIC allows us to select training data as effectively as typical active learning, without the cost of manual annotation. Finally, we argue that the speed-up achieved when learning actively in this social context (where labels can be obtained without the cost of human annotation) is necessary to cope with the continuously growing requirements of large-scale applications. In this respect, we demonstrate that SALIC requires ten times less training data in order to reach the same performance as a straightforward informativeness-agnostic learning approach.
引用
收藏
页码:1488 / 1503
页数:16
相关论文
共 40 条
  • [11] Cui P, 2016, IEEE MULTIMEDIA, V23, P92, DOI 10.1109/MMUL.2016.8
  • [12] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [13] Ebert S, 2012, PROC CVPR IEEE, P3626, DOI 10.1109/CVPR.2012.6248108
  • [14] Fang M, 2014, AAAI CONF ARTIF INTE, P1809
  • [15] Freytag A, 2014, LECT NOTES COMPUT SC, V8692, P562, DOI 10.1007/978-3-319-10593-2_37
  • [16] Relaxing from Vocabulary: Robust Weakly-Supervised Deep Learning for Vocabulary-Free Image Tagging
    Fu, Jianlong
    Wu, Yue
    Mei, Tao
    Wang, Jinqiao
    Lu, Hanqing
    Rui, Yong
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1985 - 1993
  • [17] Golge E, 2014, LECT NOTES COMPUT SC, V8695, P439, DOI 10.1007/978-3-319-10584-0_29
  • [18] Hoi SCH, 2005, PROC CVPR IEEE, P302
  • [19] Deep Classifiers from Image Tags in the Wild
    Izadinia, Hamid
    Russell, Bryan C.
    Farhadi, Ali
    Hoffman, Matthew D.
    Hertzmann, Aaron
    [J]. MMCOMMONS'15: PROCEEDINGS OF THE 2015 WORKSHOP ON COMMUNITY-ORGANIZED MULTIMODAL MINING: OPPORTUNITIES FOR NOVEL SOLUTIONS, 2015, : 13 - 18
  • [20] Joachims Thorsten, 2005, P 10 EUR C MACH LEAR, P137, DOI DOI 10.1007/BFB0026683