PROMPTING TO PROMPT FOR REHEARSAL-FREE CLASS INCREMENTAL LEARNING

被引:0
作者
Zhao, Guangzhi [1 ]
Hou, Yuting [1 ]
Mu, Kedian [1 ]
机构
[1] Peking Univ, Sch Math Sci, Beijing, Peoples R China
来源
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024 | 2024年
基金
中国国家自然科学基金;
关键词
Prompt; Catastrophic Forgetting; Class Incremental Learning; Vision Transformer;
D O I
10.1109/ICASSP48485.2024.10447170
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Class incremental learning (CIL) devotes to addressing catastrophic forgetting while continually learning new tasks. Recently, prompt tuning techniques based on vision transformers (ViT) have achieved promising results in rehearsal-free CIL. To alleviate forgetting, representative methods use a query-key mechanism to generate prompts and attach them to the frozen pre-trained ViT. However, these methods neglect the effect of query, and the learning capacity of the model is limited due to unsuitable prompts. In this paper, we propose a new approach called Prompting to Prompt (P2P). Instead of using a task-independent query function, we learn sample queries together with prompts in response to the shift of data distribution in CIL. P2P can better separate classes across tasks because the generated prompts are effective and more discriminative sample features can be extracted. Besides, the whole training process is end-to-end and queries are decided by prompts themselves, which avoids additional parameters. P2P improves the plasticity of model while maintaining good resistance to forgetting in the long task sequence. Experiments show that our approach achieves state-of-the-art results with even fewer parameters.
引用
收藏
页码:5405 / 5409
页数:5
相关论文
共 19 条
[1]  
Chaudhry A., 2018, arXiv
[2]  
Dosovitskiy A., 2020, arXiv, V2010, P11929, DOI [10.48550/arXiv.2010.11929Focustolearnmore, DOI 10.48550/ARXIV.2010.11929FOCUSTOLEARNMORE]
[3]  
Fernando Chrisantha, 2017, arXiv preprint arXiv:1701.08734
[4]  
Finn C, 2017, PR MACH LEARN RES, V70
[5]   The Many Faces of Robustness: A Critical Analysis of Out-of-Distribution Generalization [J].
Hendrycks, Dan ;
Basart, Steven ;
Mu, Norman ;
Kadavath, Saurav ;
Wang, Frank ;
Dorundo, Evan ;
Desai, Rahul ;
Zhu, Tyler ;
Parajuli, Samyak ;
Guo, Mike ;
Song, Dawn ;
Steinhardt, Jacob ;
Gilmer, Justin .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :8320-8329
[6]  
Hu WP, 2021, AAAI CONF ARTIF INTE, V35, P7797
[7]   Visual Prompt Tuning [J].
Jia, Menglin ;
Tang, Luming ;
Chen, Bor-Chun ;
Cardie, Claire ;
Belongie, Serge ;
Hariharan, Bharath ;
Lim, Ser-Nam .
COMPUTER VISION - ECCV 2022, PT XXXIII, 2022, 13693 :709-727
[8]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001
[9]   Overcoming catastrophic forgetting in neural networks [J].
Kirkpatricka, James ;
Pascanu, Razvan ;
Rabinowitz, Neil ;
Veness, Joel ;
Desjardins, Guillaume ;
Rusu, Andrei A. ;
Milan, Kieran ;
Quan, John ;
Ramalho, Tiago ;
Grabska-Barwinska, Agnieszka ;
Hassabis, Demis ;
Clopath, Claudia ;
Kumaran, Dharshan ;
Hadsell, Raia .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2017, 114 (13) :3521-3526
[10]  
Krizhevsky A., 2009, Technical report