Rethinking Few-Shot Class-Incremental Learning With Open-Set Hypothesis in Hyperbolic Geometry

被引:1
作者
Cui, Yawen [1 ]
Yu, Zitong [2 ]
Peng, Wei [3 ]
Tian, Qi [4 ]
Liu, Li [5 ,6 ]
机构
[1] Univ Oulu, CMVS, Oulu 90570, Finland
[2] Great Bay Univ, Dongguan 523000, Guangdong, Peoples R China
[3] Stanford Univ, CNSlab, Stanford, CA 94305 USA
[4] Xidian Univ, Xian 710071, Peoples R China
[5] Natl Univ Def Technol NUDT, Coll Elect Sci, Changsha 410073, Peoples R China
[6] Univ Oulu, Ctr Machine Vis & Signal Anal CMVS, Oulu 90570, Finland
关键词
Few-shot learning; class-incremental learning; hyperbolic deep neural network; open-set recognition; knowledge distillation;
D O I
10.1109/TMM.2023.3340550
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
By training first with a large base dataset, Few-Shot Class-Incremental Learning (FSCIL) aims at continually learning a sequence of few-shot learning tasks with novel classes. There are mainly two challenges in FSCIL: the overfitting issue of novel classes with limited labeled samples and the catastrophic forgetting of previously seen classes. The current protocol of FSCIL is built by mimicking the general class-incremental learning setting by building a unified framework, while the existing frameworks for FSCIL on this protocol always bias to the classes in the base dataset because the dominant performance of the deep model is decided by the size of the training dataset. Moreover, it is difficult to handle the stability-plasticity constraint in a unified FSCIL framework. To solve these issues, we rethink the configuration of FSCIL with the open-set hypothesis by reserving the possibility in the first session for incoming categories. To find a better decision boundary of close space and open space, Hyperbolic Reciprocal Point Learning module (Hyper-RPL) is built on Reciprocal Point Learning with hyperbolic neural networks. Besides, when learning novel categories from limited labeled data, we incorporate a hyperbolic metric learning (Hyper-Metric) module into the distillation-based framework to alleviate the overfitting issue and better handle the trade-off issue between the preservation of old knowledge and the acquisition of new knowledge. Finally, the comprehensive assessments of the proposed configuration and modules on three benchmark datasets are executed to validate the effectiveness, and state-of-the-art results are achieved.
引用
收藏
页码:5897 / 5910
页数:14
相关论文
共 67 条
[1]   Few-Shot Class Incremental Learning Leveraging Self-Supervised Features [J].
Ahmad, Touqeer ;
Dhamija, Akshay Raj ;
Cruz, Steve ;
Rabinowitz, Ryan ;
Li, Chunchun ;
Jafarzadeh, Mohsen ;
Boult, Terrance E. .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, :3899-3909
[2]  
Aly R, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P4811
[3]  
Bachmann G, 2020, PR MACH LEARN RES, V119
[4]   Towards Open Set Deep Networks [J].
Bendale, Abhijit ;
Boult, Terrance E. .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1563-1572
[5]  
Cannon J. W., 1997, Flavors of geometry, V31, P2
[6]   End-to-End Incremental Learning [J].
Castro, Francisco M. ;
Marin-Jimenez, Manuel J. ;
Guil, Nicolas ;
Schmid, Cordelia ;
Alahari, Karteek .
COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 :241-257
[7]   A Comprehensive Survey of Scene Graphs: Generation and Application [J].
Chang, Xiaojun ;
Ren, Pengzhen ;
Xu, Pengfei ;
Li, Zhihui ;
Chen, Xiaojiang ;
Hauptmann, Alex .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (01) :1-26
[8]  
Chaudhry A., 2018, PROC INT C LEARN RE
[9]   Learning Open Set Network with Discriminative Reciprocal Points [J].
Chen, Guangyao ;
Qiao, Limeng ;
Shi, Yemin ;
Peng, Peixi ;
Li, Jia ;
Huang, Tiejun ;
Pu, Shiliang ;
Tian, Yonghong .
COMPUTER VISION - ECCV 2020, PT III, 2020, 12348 :507-522
[10]   Dual-Awareness Attention for Few-Shot Object Detection [J].
Chen, Tung-, I ;
Liu, Yueh-Cheng ;
Su, Hung-Ting ;
Chang, Yu-Cheng ;
Lin, Yu-Hsiang ;
Yeh, Jia-Fong ;
Chen, Wen-Chin ;
Hsu, Winston H. .
IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 :291-301