Cold-start active learning for image classification

被引:23
作者
Jin, Qiuye [1 ,2 ]
Yuan, Mingzhi [1 ,2 ]
Li, Shiman [1 ,2 ]
Wang, Haoran [1 ,2 ]
Wang, Manning [1 ,2 ]
Song, Zhijian [1 ,2 ]
机构
[1] Fudan Univ, Digital Med Res Ctr, Sch Basic Med Sci, Shanghai 200032, Peoples R China
[2] Shanghai Key Lab Med Image Comp & Comp Assisted In, Shanghai 200032, Peoples R China
基金
中国国家自然科学基金;
关键词
Active learning; Cold start; Image classification; DENSITY;
D O I
10.1016/j.ins.2022.10.066
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Active learning (AL) aims to select valuable samples for labeling from an unlabeled sample pool to build a training dataset with minimal annotation cost. Traditional methods always require partially and initially labeled samples to start active selection and then query annotations of samples incrementally through several iterations. However, this scheme is not effective in the deep learning scenario. On the one hand, initially labeled sample sets are not always available in the beginning. On the other hand, the performance of the traditional model is usually poor in the early iterations due to limited training feedback. For the first time, we propose a cold-start AL model based on representative (CALR) sampling, which selects valuable samples without the need for an initial labeled set or the iterative feedback of the target models. Experiments on three image classification datasets, CIFAR-10, CIFAR-100 and Caltech-256, showed that CALR achieved a new state-of-the-art performance of AL in cold-start settings. Especially in low annotation budget conditions, our method can achieve up to a 10% performance increase compared to traditional methods. Furthermore, CALR can be combined with warm-start methods to improve the start-up efficiency while further breaking the performance ceiling of AL, which makes CALR have a broader application scenario. (C) 2022 Elsevier Inc. All rights reserved.
引用
收藏
页码:16 / 36
页数:21
相关论文
共 48 条
[1]  
[Anonymous], 2009, Rep. TR-2009
[2]  
[Anonymous], 1998, P 15 INT C MACH LEAR
[3]   The power of ensembles for active learning in image classification [J].
Beluch, William H. ;
Genewein, Tim ;
Nuernberger, Andreas ;
Koehler, Jan M. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9368-9377
[4]   Ranked batch-mode active learning [J].
Cardoso, Thiago N. C. ;
Silva, Rodrigo M. ;
Canuto, Sergio ;
Moro, Mirella M. ;
Goncalves, Marcos A. .
INFORMATION SCIENCES, 2017, 379 :313-337
[5]   Active learning Bayesian support vector regression model for global approximation [J].
Cheng, Kai ;
Lu, Zhenzhou .
INFORMATION SCIENCES, 2021, 544 :549-563
[6]  
COHN D, 1994, MACH LEARN, V15, P201, DOI 10.1007/BF00993277
[7]  
Danka T, 2018, Arxiv, DOI [arXiv:1805.00979, DOI 10.48550/ARXIV.1805.00979, 10.48550/arXiv.1805.00979]
[8]   Broad and deep neural network for high-dimensional data representation learning [J].
Feng, Qiying ;
Liu, Zhulin ;
Chen, C. L. Philip .
INFORMATION SCIENCES, 2022, 599 :127-146
[9]  
Freytag A, 2014, LECT NOTES COMPUT SC, V8692, P562, DOI 10.1007/978-3-319-10593-2_37
[10]  
Gidaris S, 2018, Arxiv, DOI arXiv:1803.07728