Intra-Class Adaptive Augmentation With Neighbor Correction for Deep Metric Learning

被引:2
作者
Fu, Zheren [1 ]
Mao, Zhendong [2 ,3 ]
Hu, Bo [2 ]
Liu, An-An [4 ]
Zhang, Yongdong [2 ,3 ]
机构
[1] Univ Sci & Technol China, Sch Cyberspace Sci & Technol, Hefei 230027, Peoples R China
[2] Univ Sci & Technol China, Sch Informat Sci & Technol, Hefei 230022, Peoples R China
[3] Hefei Comprehens Natl Sci Ctr, Inst Artificial Intelligence, Hefei 230022, Peoples R China
[4] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep metric learning; distribution estimation; sample generation; semantic augmentation;
D O I
10.1109/TMM.2022.3227414
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep metric learning aims to learn an embedding space, where semantically similar samples are close together and dissimilar ones are repelled against. To explore more hard and informative training signals for augmentation and generalization, recent methods focus on generating synthetic samples to boost metric learning losses. However, these methods just use the deterministic and class-independent generations (e.g., simple linear interpolation), which only can cover the limited part of distribution spaces around original samples. They have overlooked the wide characteristic changes of different classes and can not model abundant intra-class variations for generations. Therefore, generated samples not only lack rich semantics within the certain class, but also might be noisy signals to disturb training. In this paper, we propose a novel intra-class adaptive augmentation (IAA) framework for deep metric learning. We reasonably estimate intra-class variations for every class and generate adaptive synthetic samples to support hard samples mining and boostmetric learning losses. Further, for most datasets that have a few samples within the class, we propose the neighbor correction to revise the inaccurate estimations, according to our correlation discovery where similar classes generally have similar variation distributions. Extensive experiments on five benchmarks show our method significantly improves and outperforms the state-of-the-art methods on retrieval performances by 3%-6%.
引用
收藏
页码:7758 / 7771
页数:14
相关论文
共 75 条
  • [21] Ioffe S, 2015, PR MACH LEARN RES, V37, P448
  • [22] Deep Metric Learning With Manifold Class Variability Analysis
    Kim, Dae Ha
    Song, Byung Cheol
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 24 : 3533 - 3544
  • [23] Proxy Anchor Loss for Deep Metric Learning
    Kim, Sungyeon
    Kim, Dongwon
    Cho, Minsu
    Kwak, Suha
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 3235 - 3244
  • [24] Embedding Expansion: Augmentation in Embedding Space for Deep Metric Learning
    Ko, Byungsoo
    Gu, Geonmo
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 7253 - 7262
  • [25] 3D Object Representations for Fine-Grained Categorization
    Krause, Jonathan
    Stark, Michael
    Deng, Jia
    Li Fei-Fei
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2013, : 554 - 561
  • [26] Krizhevsky A., 2009, Technical Report 0
  • [27] Lee K., 2022, PROC BRIT MACH VIS
  • [28] Learning generative visual models from few training examples: An incremental Bayesian approach tested on 101 object categories
    Li Fei-Fei
    Fergus, Rob
    Perona, Pietro
    [J]. COMPUTER VISION AND IMAGE UNDERSTANDING, 2007, 106 (01) : 59 - 70
  • [29] MetaSAug: Meta Semantic Augmentation for Long-Tailed Visual Recognition
    Li, Shuang
    Gong, Kaixiong
    Liu, Chi Harold
    Wang, Yulin
    Qiao, Feng
    Cheng, Xinjing
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 5208 - 5217
  • [30] Deep Variational Metric Learning
    Lin, Xudong
    Duan, Yueqi
    Dong, Qiyuan
    Lu, Jiwen
    Zhou, Jie
    [J]. COMPUTER VISION - ECCV 2018, PT 15, 2018, 11219 : 714 - 729