Meta label associated loss for fine-grained visual recognition

被引:0
|
作者
Li, Yanchao [1 ]
Xiao, Fu [1 ]
Li, Hao [2 ]
Li, Qun [1 ]
Yu, Shui [3 ]
机构
[1] Nanjing Univ Posts & Telecommun, Sch Comp Sci, Nanjing 210023, Peoples R China
[2] Zhoukou Normal Univ, Sch Network Engn, Zhoukou 466001, Peoples R China
[3] Univ Technol Sydney, Sch Comp Sci, Sydney 2007, Australia
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
label associated loss; weighting noisy samples; fine-grained visual recognition; noise-tolerant learning; meta-learning;
D O I
10.1007/s11432-023-3922-2
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, intensive attempts have been made to design robust models for fine-grained visual recognition, most notably are the impressive gains for training with noisy labels by incorporating a reweighting strategy into a meta-learning framework. However, it is limited to up or downweighting the contribution of an instance for label reweighting approaches in the learning process. To solve this issue, a novel noise-tolerant method with auxiliary web data is proposed. Specifically, first, the associations made from embeddings of well-labeled data with those of web data and back at the same class are measured. Next, its association probability is employed as a weighting fusion strategy into angular margin-based loss, which makes the trained model robust to noisy datasets. To reduce the influence of the gap between the well-labeled and noisy web data, a bridge schema is proposed via the corresponding loss that encourages the learned embeddings to be coherent. Lastly, the formulation is encapsulated into the meta-learning framework, which can reduce the overfitting of models and learn the network parameters to be noise-tolerant. Extensive experiments are performed on benchmark datasets, and the results clearly show the superiority of the proposed method over existing state-of-the-art approaches.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Meta label associated loss for fine-grained visual recognition
    Yanchao LI
    Fu XIAO
    Hao LI
    Qun LI
    Shui YU
    ScienceChina(InformationSciences), 2024, 67 (06) : 230 - 247
  • [2] Robust Fine-Grained Visual Recognition With Neighbor-Attention Label Correction
    Mao, Shunan
    Zhang, Shiliang
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 2614 - 2626
  • [3] Fine-Grained Crowdsourcing for Fine-Grained Recognition
    Jia Deng
    Krause, Jonathan
    Li Fei-Fei
    2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 580 - 587
  • [4] Annotation modification for fine-grained visual recognition
    Luo, Changzhi
    Meng, Zhijun
    Feng, Jiashi
    Ni, Bingbing
    Wang, Meng
    NEUROCOMPUTING, 2018, 274 : 58 - 65
  • [5] ProtoSimi: label correction for fine-grained visual categorization
    Shen, Jialiang
    Yao, Yu
    Huang, Shaoli
    Wang, Zhiyong
    Zhang, Jing
    Wang, Ruxing
    Yu, Jun
    Liu, Tongliang
    MACHINE LEARNING, 2024, 113 (04) : 1903 - 1920
  • [6] ProtoSimi: label correction for fine-grained visual categorization
    Jialiang Shen
    Yu Yao
    Shaoli Huang
    Zhiyong Wang
    Jing Zhang
    Ruxing Wang
    Jun Yu
    Tongliang Liu
    Machine Learning, 2024, 113 : 1903 - 1920
  • [7] Bilinear CNN Models for Fine-grained Visual Recognition
    Lin, Tsung-Yu
    RoyChowdhury, Aruni
    Maji, Subhransu
    2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1449 - 1457
  • [8] Hierarchical Bilinear Pooling for Fine-Grained Visual Recognition
    Yu, Chaojian
    Zhao, Xinyi
    Zheng, Qi
    Zhang, Peng
    You, Xinge
    COMPUTER VISION - ECCV 2018, PT XVI, 2018, 11220 : 595 - 610
  • [9] Fine-grained object recognition in underwater visual data
    Spampinato, C.
    Palazzo, S.
    Joalland, P. H.
    Paris, S.
    Glotin, H.
    Blanc, K.
    Lingrand, D.
    Precioso, F.
    MULTIMEDIA TOOLS AND APPLICATIONS, 2016, 75 (03) : 1701 - 1720
  • [10] Fine-grained object recognition in underwater visual data
    C. Spampinato
    S. Palazzo
    P. H. Joalland
    S. Paris
    H. Glotin
    K. Blanc
    D. Lingrand
    F. Precioso
    Multimedia Tools and Applications, 2016, 75 : 1701 - 1720