Exploration of Class Center for Fine-Grained Visual Classification

被引:5
作者
Yao, Hang [1 ,2 ]
Miao, Qiguang [1 ,2 ]
Zhao, Peipei [1 ,2 ]
Li, Chaoneng [1 ,2 ]
Li, Xin [3 ]
Feng, Guanwen [1 ,2 ]
Liu, Ruyi
机构
[1] Xidian Univ, Sch Comp Sci & Technol, Xian Key Lab Big Data & Intelligent Vis, Xian 710071, Shaanxi, Peoples R China
[2] Xidian Univ, Key Lab Collaborat Intelligence Syst, Minist Educ, Xian 710071, Shaanxi, Peoples R China
[3] Yanshan Univ, Sch Mech Engn, Qinhuangdao 066004, Peoples R China
关键词
Visualization; Feature extraction; Predictive models; Task analysis; Training; Reliability; Optimization; Fine-grained visual classification; exploration of class center; class center; soft label;
D O I
10.1109/TCSVT.2024.3406443
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Different from large-scale classification tasks, fine-grained visual classification is a challenging task due to two critical problems: 1) evident intra-class variances and subtle inter-class differences, and 2) overfitting owing to fewer training samples in datasets. Most existing methods extract key features to reduce intra-class variances, but pay no attention to subtle inter-class differences in fine-grained visual classification. To address this issue, we propose a loss function named exploration of class center, which consists of a multiple class-center constraint and a class-center label generation. This loss function fully utilizes the information of the class center from the perspective of features and labels. From the feature perspective, the multiple class-center constraint pulls samples closer to the target class center, and pushes samples away from the most similar nontarget class center. Thus, the constraint reduces intra-class variances and enlarges inter-class differences. From the label perspective, the class-center label generation utilizes class-center distributions to generate soft labels to alleviate overfitting. Our method can be easily integrated with existing fine-grained visual classification approaches as a loss function, to further boost excellent performance with only slight training costs. Extensive experiments are conducted to demonstrate consistent improvements achieved by our method on four widely-used fine-grained visual classification datasets. In particular, our method achieves state-of-the-art performance on the FGVC-Aircraft and CUB-200-2011 datasets.
引用
收藏
页码:9954 / 9966
页数:13
相关论文
共 68 条
[11]  
Hadsell R., 2006, P IEEE COMP SOC C CO, V2, P1735
[12]  
He J, 2021, Arxiv, DOI arXiv:2103.07976
[13]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[14]   Which and How Many Regions to Gaze: Focus Discriminative Regions for Fine-Grained Visual Categorization [J].
He, Xiangteng ;
Peng, Yuxin ;
Zhao, Junjie .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2019, 127 (09) :1235-1255
[15]  
Hinton G, 2015, Arxiv, DOI arXiv:1503.02531
[16]  
Hu T, 2019, Arxiv, DOI [arXiv:1901.09891, DOI 10.48550/ARXIV.1901.09891]
[17]   Densely Connected Convolutional Networks [J].
Huang, Gao ;
Liu, Zhuang ;
van der Maaten, Laurens ;
Weinberger, Kilian Q. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2261-2269
[18]   Local descriptor-based multi-prototype network for few-shot Learning [J].
Huang, Hongwei ;
Wu, Zhangkai ;
Li, Wenbin ;
Huo, Jing ;
Gao, Yang .
PATTERN RECOGNITION, 2021, 116
[19]   Part-Stacked CNN for Fine-Grained Visual Categorization [J].
Huang, Shaoli ;
Xu, Zhe ;
Tao, Dacheng ;
Zhang, Ya .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1173-1182
[20]   Interpretable and Accurate Fine-grained Recognition via Region Grouping [J].
Huang, Zixuan ;
Li, Yin .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :8659-8669