Multi-Granularity Regularized Re-Balancing for Class Incremental Learning

被引:9
作者
Chen, Huitong [1 ,2 ]
Wang, Yu [2 ,3 ,4 ]
Hu, Qinghua [2 ,3 ,4 ]
机构
[1] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300350, Peoples R China
[2] Key Lab Machine Learning Tianjin, Tianjin 300350, Peoples R China
[3] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300350, Peoples R China
[4] Haihe Lab Informat Technol Applicat Innova t, Tianjin 300020, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Class hierarchy; class incremental learning; multi-granularity regularization; re-balancing modeling;
D O I
10.1109/TKDE.2022.3188335
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning models suffer from catastrophic forgetting when learning new tasks incrementally. Incremental learning has been proposed to retain the knowledge of old classes while learning to identify new classes. A typical approach is to use a few exemplars to avoid forgetting old knowledge. In such a scenario, data imbalance between old and new classes is a key issue that leads to performance degradation of the model. Several strategies have been designed to rectify the bias towards the new classes due to data imbalance. However, they heavily rely on the assumptions of the bias relation between old and new classes. Therefore, they are not suitable for complex real-world applications. In this study, we propose an assumption-agnostic method, Multi-Granularity Regularized re-Balancing (MGRB), to address this problem. Re-balancing methods are used to alleviate the influence of data imbalance; however, we empirically discover that they would under-fit new classes. To this end, we further design a novel multi-granularity regularization term that enables the model to consider the correlations of classes in addition to re-balancing the data. A class hierarchy is first constructed by ontology or grouping semantically or visually similar classes. The multi-granularity regularization then transforms the one-hot label vector into a continuous label distribution, which reflects the relations between the target class and other classes based on the constructed class hierarchy. Thus, the model can learn the inter-class relational information, which helps enhance the learning of both old and new classes. Experimental results on both public datasets and a real-world fault diagnosis dataset verify the effectiveness of the proposed method. Code is available at https://github.com/lilyht/CIL-MGRB.
引用
收藏
页码:7263 / 7277
页数:15
相关论文
共 51 条
  • [1] Rainbow Memory: Continual Learning with a Memory of Diverse Samples
    Bang, Jihwan
    Kim, Heesu
    Yoo, YoungJoon
    Ha, Jung-Woo
    Choi, Jonghyun
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 8214 - 8223
  • [2] IL2M: Class Incremental Learning With Dual Memory
    Belouadah, Eden
    Popescu, Adrian
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 583 - 592
  • [3] Bengio S, 2010, NIPS, V23, P163
  • [4] Bertinetto L, 2020, PROC CVPR IEEE, P12503, DOI 10.1109/CVPR42600.2020.01252
  • [5] End-to-End Incremental Learning
    Castro, Francisco M.
    Marin-Jimenez, Manuel J.
    Guil, Nicolas
    Schmid, Cordelia
    Alahari, Karteek
    [J]. COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 : 241 - 257
  • [6] Chaudhry A, 2021, Arxiv, DOI arXiv:2002.08165
  • [7] SMOTE: Synthetic minority over-sampling technique
    Chawla, Nitesh V.
    Bowyer, Kevin W.
    Hall, Lawrence O.
    Kegelmeyer, W. Philip
    [J]. 2002, American Association for Artificial Intelligence (16)
  • [8] Class-Balanced Loss Based on Effective Number of Samples
    Cui, Yin
    Jia, Menglin
    Lin, Tsung-Yi
    Song, Yang
    Belongie, Serge
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 9260 - 9269
  • [9] Douillard Arthur, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12365), P86, DOI 10.1007/978-3-030-58565-5_6
  • [10] ADASYN: Adaptive Synthetic Sampling Approach for Imbalanced Learning
    He, Haibo
    Bai, Yang
    Garcia, Edwardo A.
    Li, Shutao
    [J]. 2008 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-8, 2008, : 1322 - 1328