Understanding the Impact of Label Granularity on CNN-based Image Classification

被引:14
作者
Chen, Zhuo [1 ]
Ding, Ruizhou [1 ]
Chin, Ting-Wu [1 ]
Marculescu, Diana [1 ]
机构
[1] Carnegie Mellon Univ, Elect & Comp Engn, Pittsburgh, PA 15213 USA
来源
2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW) | 2018年
关键词
Convolutional Neural Networks; Supervised Learning; Image Classification; Labeling;
D O I
10.1109/ICDMW.2018.00131
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, supervised learning using Convolutional Neural Networks (CNNs) has achieved great success in image classification tasks, and large scale labeled datasets have contributed significantly to this achievement. However, the definition of a label is often application dependent. For example, an image of a cat can be labeled as "cat" or perhaps more specifically "Persian cat." We refer to this as label granularity. In this paper, we conduct extensive experiments using various datasets to demonstrate and analyze how and why training based on fine-grain labeling, such as "Persian cat" can improve CNN accuracy on classifying coarse-grain classes, in this case "cat." The experimental results show that training CNNs with fine-grain labels improves both network's optimization and generalization capabilities, as intuitively it encourages the network to learn more features, and hence increases classification accuracy on coarse-grain classes under all datasets considered. Moreover, fine-grain labels enhance data efficiency in CNN training. For example, a CNN trained with fine-grain labels and only 40% of the total training data can achieve higher accuracy than a CNN trained with the full training dataset and coarse-grain labels. These results point to two possible applications of this work: (i) with sufficient human resources, one can improve CNN performance by re-labeling the dataset with fine-grain labels, and (ii) with limited human resources, to improve CNN performance, rather than collecting more training data, one may instead use fine-grain labels for the dataset. We also observe that the improvement brought by fine-grain labeling varies from dataset to dataset, therefore we further propose a metric called Average Confusion Ratio to characterize the effectiveness of fine-grain labeling, and show its use through extensive experimentation. Code is available at https://github.com/cmu-enyac/Label-Granularity.
引用
收藏
页码:895 / 904
页数:10
相关论文
共 33 条
[1]   On estimating simple probabilistic discriminative models with subclasses [J].
Ahmed, Nisar ;
Campbell, Mark .
EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (07) :6659-6664
[2]  
[Anonymous], 2004, PROC INT C MACH LEAR
[3]  
[Anonymous], 2014, AAAI
[4]   Hierarchical Multilabel Classification with Minimum Bayes Risk [J].
Bi, Wei ;
Kwok, James T. .
12TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2012), 2012, :101-110
[5]   Weakly Supervised Deep Detection Networks [J].
Bilen, Hakan ;
Vedaldi, Andrea .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2846-2854
[6]  
Cerri R, 2015, IEEE IJCNN
[7]  
Cesa-Bianchi N, 2006, J MACH LEARN RES, V7, P31
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]   Clustering Inside Classes Improves Performance of Linear Classifiers [J].
Fradkin, Dmitriy .
20TH IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, VOL 2, PROCEEDINGS, 2008, :439-442
[10]  
Glorot X, 2010, P 13 INT C ART INT S, P249, DOI DOI 10.1109/LGRS.2016.2565705