Beyond Max-Margin: Class Margin Equilibrium for Few-shot Object Detection

被引:117
作者
Li, Bohao [1 ]
Yang, Boyu [1 ]
Liu, Chang [1 ]
Liu, Feng [1 ]
Ji, Rongrong [2 ,3 ,4 ]
Ye, Qixiang [1 ,4 ]
机构
[1] Univ Chinese Acad Sci, EECE, PriSDL, Beijing 100049, Peoples R China
[2] Xiamen Univ, MAC, Dept Artificial Intelligence, Sch Informat, Xiamen 361005, Peoples R China
[3] Xiamen Univ, Inst Artificial Intelligence, Xiamen 361005, Peoples R China
[4] Peng Cheng Lab, Shenzhen, Peoples R China
来源
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021 | 2021年
关键词
D O I
10.1109/CVPR46437.2021.00728
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot object detection has made substantial progress by representing novel class objects using the feature representation learned upon a set of base class objects. However, an implicit contradiction between novel class classification and representation is unfortunately ignored. On the one hand, to achieve accurate novel class classification, the distributions of either two base classes must be far away from each other (max-margin). On the other hand, to precisely represent novel classes, the distributions of base classes should be close to each other to reduce the intra-class distance of novel classes (min-margin). In this paper, we propose a class margin equilibrium (CME) approach, with the aim to optimize both feature space partition and novel class reconstruction in a systematic way. CME first converts the few-shot detection problem to the few-shot classification problem by using a fully connected layer to decouple localization features. CME then reserves adequate margin space for novel classes by introducing simple-yet-effective class margin loss during feature learning. Finally, CME pursues margin equilibrium by disturbing the features of novel class instances in an adversarial min-max fashion. Experiments on Pascal VOC and MS-COCO datasets show that CME significantly improves upon two baseline detectors (up to 3 similar to 5% in average), achieving state-of-the-art performance.
引用
收藏
页码:7359 / 7368
页数:10
相关论文
共 34 条
[1]  
Chen H, 2018, AAAI CONF ARTIF INTE, P2836
[2]  
Finn C, 2017, PR MACH LEARN RES, V70
[3]   Fast R-CNN [J].
Girshick, Ross .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1440-1448
[4]   Collect and Select: Semantic Alignment Metric Learning for Few-Shot Learning [J].
Hao, Fusheng ;
He, Fengxiang ;
Cheng, Jun ;
Wang, Lei ;
Cao, Jianzhong ;
Tao, Dacheng .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :8459-8468
[5]   Low-shot Visual Recognition by Shrinking and Hallucinating Features [J].
Hariharan, Bharath ;
Girshick, Ross .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :3037-3046
[6]   Task Agnostic Meta-Learning for Few-Shot Learning [J].
Jamal, Muhammad Abdullah ;
Qi, Guo-Jun .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11711-11719
[7]   Few-shot Object Detection via Feature Reweighting [J].
Kang, Bingyi ;
Liu, Zhuang ;
Wang, Xin ;
Yu, Fisher ;
Feng, Jiashi ;
Darrell, Trevor .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :8419-8428
[8]   Boosting Few-Shot Learning with Adaptive Margin Loss [J].
Li, Aoxue ;
Huang, Weiran ;
Lan, Xu ;
Feng, Jiashi ;
Li, Zhenguo ;
Wang, Liwei .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :12573-12581
[9]  
Li X., 2020, REMOTE SENSING, V12
[10]   Feature Pyramid Networks for Object Detection [J].
Lin, Tsung-Yi ;
Dollar, Piotr ;
Girshick, Ross ;
He, Kaiming ;
Hariharan, Bharath ;
Belongie, Serge .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :936-944