Exploration of Class Center for Fine-Grained Visual Classification

被引:5
作者
Yao, Hang [1 ,2 ]
Miao, Qiguang [1 ,2 ]
Zhao, Peipei [1 ,2 ]
Li, Chaoneng [1 ,2 ]
Li, Xin [3 ]
Feng, Guanwen [1 ,2 ]
Liu, Ruyi
机构
[1] Xidian Univ, Sch Comp Sci & Technol, Xian Key Lab Big Data & Intelligent Vis, Xian 710071, Shaanxi, Peoples R China
[2] Xidian Univ, Key Lab Collaborat Intelligence Syst, Minist Educ, Xian 710071, Shaanxi, Peoples R China
[3] Yanshan Univ, Sch Mech Engn, Qinhuangdao 066004, Peoples R China
关键词
Visualization; Feature extraction; Predictive models; Task analysis; Training; Reliability; Optimization; Fine-grained visual classification; exploration of class center; class center; soft label;
D O I
10.1109/TCSVT.2024.3406443
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Different from large-scale classification tasks, fine-grained visual classification is a challenging task due to two critical problems: 1) evident intra-class variances and subtle inter-class differences, and 2) overfitting owing to fewer training samples in datasets. Most existing methods extract key features to reduce intra-class variances, but pay no attention to subtle inter-class differences in fine-grained visual classification. To address this issue, we propose a loss function named exploration of class center, which consists of a multiple class-center constraint and a class-center label generation. This loss function fully utilizes the information of the class center from the perspective of features and labels. From the feature perspective, the multiple class-center constraint pulls samples closer to the target class center, and pushes samples away from the most similar nontarget class center. Thus, the constraint reduces intra-class variances and enlarges inter-class differences. From the label perspective, the class-center label generation utilizes class-center distributions to generate soft labels to alleviate overfitting. Our method can be easily integrated with existing fine-grained visual classification approaches as a loss function, to further boost excellent performance with only slight training costs. Extensive experiments are conducted to demonstrate consistent improvements achieved by our method on four widely-used fine-grained visual classification datasets. In particular, our method achieves state-of-the-art performance on the FGVC-Aircraft and CUB-200-2011 datasets.
引用
收藏
页码:9954 / 9966
页数:13
相关论文
共 68 条
[1]  
[Anonymous], 2018, iNaturalist 2018 competition dataset
[2]  
Bagherinezhad H, 2018, Arxiv, DOI [arXiv:1805.02641, 10.48550/arxiv.1805.02641]
[3]  
Branson S., 2014, P BRIT MACH VIS C, DOI 10.5244/C.28.87
[4]  
Chen HX, 2023, Arxiv, DOI arXiv:2109.12932
[5]   Destruction and Construction Learning for Fine-grained Image Recognition [J].
Chen, Yue ;
Bai, Yalong ;
Zhang, Wei ;
Mei, Tao .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :5152-5161
[6]  
Dosovitskiy A., 2021, 9 INT C LEARN REPR I
[7]   Exploiting Category Similarity-Based Distributed Labeling for Fine-Grained Visual Classification [J].
Du, Pengzhen ;
Sun, Zeren ;
Yao, Yazhou ;
Tang, Zhenmin .
IEEE ACCESS, 2020, 8 :186679-186690
[8]   Fine-Grained Visual Classification via Progressive Multi-granularity Training of Jigsaw Patches [J].
Du, Ruoyi ;
Chang, Dongliang ;
Bhunia, Ayan Kumar ;
Xie, Jiyang ;
Ma, Zhanyu ;
Song, Yi-Zhe ;
Guo, Jun .
COMPUTER VISION - ECCV 2020, PT XX, 2020, 12365 :153-168
[9]   Pairwise Confusion for Fine-Grained Visual Classification [J].
Dubey, Abhimanyu ;
Gupta, Otkrist ;
Guo, Pei ;
Raskar, Ramesh ;
Farrell, Ryan ;
Naik, Nikhil .
COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 :71-88
[10]   Facial Expression Recognition in the Wild via Deep Attentive Center Loss [J].
Farzaneh, Amir Hossein ;
Qi, Xiaojun .
2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, :2401-2410