From Less to More: Progressive Generalized Zero-Shot Detection With Curriculum Learning

被引:12
作者
Liu, Jingren [1 ]
Chen, Yi [2 ]
Liu, Huajun [1 ]
Zhang, Haofeng [1 ]
Zhang, Yudong [3 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
[2] Nanjing Normal Univ, Sch Comp Sci & Technol, Nanjing 210023, Peoples R China
[3] Univ Leicester, Sch Informat, Leicester LE1 7RH, Leics, England
基金
中国国家自然科学基金;
关键词
Task analysis; Visualization; Generators; Training; Object detection; Semantics; Proposals; generalized zero-shot detection (GZSD); curriculum learning; generative adversarial network (GAN); NETWORK;
D O I
10.1109/TITS.2022.3151073
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Object detection, as one of the most important environment perception tasks for traffic safety in intelligent transportation systems, has been widely investigated recently. However, most of the researches focus on the fully supervised scenario, and inevitably lead to model failure. With the continuous development of Zero-Shot Learning (ZSL) models, Generalized Zero-Shot Detection (GZSD) has attracted great attention due to its ability of detecting unseen objects. Many researchers tend to map the detected visual features to semantic attributes and then separate seen and unseen domains during inference. But they have ignore that the generative methods generally have higher performance than these visual-semantic mapping methods, and they have been confirmed from previous GZSL methods. In order to make up for the vacancy of GZSD in the generative methods, we propose an idea of using curriculum learning to generate more precise unseen visual features. And with the excellent performance of WGAN-based method in sample synthesis, we realize the function of using semantics to generate visual features for unseen domains. In addition, we also adopt part of the idea of meta-learning to progressively correct the capability of the generator for better mitigating domain shift problem during the generation process. Through the above ideas, we can detect both seen and unseen bounding boxes and classify them accurately, by combining with the excellent detection ability of Faster-RCNN. Extensive experimental results on two popular datasets, i.e., MSCOCO and KITTI, show that our proposed method can outperform the state-of-the-art methods.
引用
收藏
页码:19016 / 19029
页数:14
相关论文
共 75 条
  • [1] Abdar M., ARXIV210508590
  • [2] A review of uncertainty quantification in deep learning: Techniques, applications and challenges
    Abdar, Moloud
    Pourpanah, Farhad
    Hussain, Sadiq
    Rezazadegan, Dana
    Liu, Li
    Ghavamzadeh, Mohammad
    Fieguth, Paul
    Cao, Xiaochun
    Khosravi, Abbas
    Acharya, U. Rajendra
    Makarenkov, Vladimir
    Nahavandi, Saeid
    [J]. INFORMATION FUSION, 2021, 76 : 243 - 297
  • [3] Label-Embedding for Attribute-Based Classification
    Akata, Zeynep
    Perronnin, Florent
    Harchaoui, Zaid
    Schmid, Cordelia
    [J]. 2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 819 - 826
  • [4] Zero-Shot Object Detection
    Bansal, Ankan
    Sikka, Karan
    Sharma, Gaurav
    Chellappa, Rama
    Divakaran, Ajay
    [J]. COMPUTER VISION - ECCV 2018, PT I, 2018, 11205 : 397 - 414
  • [5] New Tool Holder Design for Cryogenic Machining of Ti6Al4V
    Bellin, Marco
    Sartori, Stefano
    Ghiotti, Andrea
    Bruschi, Stefania
    [J]. PROCEEDINGS OF THE 20TH INTERNATIONAL ESAFORM CONFERENCE ON MATERIAL FORMING (ESAFORM 2017), 2017, 1896
  • [6] Bengio Y., 2009, P 26 ANN INT C MACHI, P41
  • [7] Berthelot D., 2017, BEGAN: Boundary Equilibrium Generative Adversarial Networks
  • [8] Carlos F., 2017, P MACHINE LEARNING R, P1
  • [9] Hybrid-Attention based Decoupled Metric Learning for Zero-Shot Image Retrieval
    Chen, Binghui
    Deng, Weihong
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2745 - 2754
  • [10] Chen Xi, 2016, ADV NEURAL INFORM PR, V29