Continual Learning From a Stream of APIs

被引:0
|
作者
Yang, Enneng [1 ]
Wang, Zhenyi [2 ]
Shen, Li [3 ,4 ]
Yin, Nan [5 ]
Liu, Tongliang [6 ]
Guo, Guibing [1 ]
Wang, Xingwei [1 ]
Tao, Dacheng [7 ]
机构
[1] Northeastern Univ, Shenyang 110004, Peoples R China
[2] Univ Maryland, College Pk, MD 20742 USA
[3] Sun Yat Sen Univ, Guangzhou 510275, Peoples R China
[4] JD Explore Acad, Beijing 101111, Peoples R China
[5] Mohamed bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
[6] Univ Sydney, Camperdown, NSW 2050, Australia
[7] Nanyang Technol Univ, Singapore 639798, Singapore
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Data-free learning; catastrophic forgetting; plasticity-stability; continual learning; NEURAL-NETWORKS; GAME; GO;
D O I
10.1109/TPAMI.2024.3460871
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continual learning (CL) aims to learn new tasks without forgetting previous tasks. However, existing CL methods require a large amount of raw data, which is often unavailable due to copyright considerations and privacy risks. Instead, stakeholders usually release pre-trained machine learning models as a service (MLaaS), which users can access via APIs. This paper considers two practical-yet-novel CL settings: data-efficient CL (DECL-APIs) and data-free CL (DFCL-APIs), which achieve CL from a stream of APIs with partial or no raw data. Performing CL under these two new settings faces several challenges: unavailable full raw data, unknown model parameters, heterogeneous models of arbitrary architecture and scale, and catastrophic forgetting of previous APIs. To overcome these issues, we propose a novel data-free cooperative continual distillation learning framework that distills knowledge from a stream of APIs into a CL model by generating pseudo data, just by querying APIs. Specifically, our framework includes two cooperative generators and one CL model, forming their training as an adversarial game. We first use the CL model and the current API as fixed discriminators to train generators via a derivative-free method. Generators adversarially generate hard and diverse synthetic data to maximize the response gap between the CL model and the API. Next, we train the CL model by minimizing the gap between the responses of the CL model and the black-box API on synthetic data, to transfer the API's knowledge to the CL model. Furthermore, we propose a new regularization term based on network similarity to prevent catastrophic forgetting of previous APIs. Our method performs comparably to classic CL with full raw data on the MNIST and SVHN datasets in the DFCL-APIs setting. In the DECL-APIs setting, our method achieves 0.97x, 0.75x and 0.69x performance of classic CL on the more challenging CIFAR10, CIFAR100, and MiniImageNet, respectively.
引用
收藏
页码:11432 / 11445
页数:14
相关论文
共 50 条
  • [31] Efficient Architecture Search for Continual Learning
    Gao, Qiang
    Luo, Zhipeng
    Klabjan, Diego
    Zhang, Fengli
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (11) : 8555 - 8565
  • [32] Continual Learning of Knowledge Graph Embeddings
    Daruna, Angel
    Gupta, Mehul
    Sridharan, Mohan
    Chernova, Sonia
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 1128 - 1135
  • [33] Example forgetting and rehearsal in continual learning
    Benko, Beatrix
    PATTERN RECOGNITION LETTERS, 2024, 179 : 65 - 72
  • [34] On Sequential Bayesian Inference for Continual Learning
    Kessler, Samuel
    Cobb, Adam
    Rudner, Tim G. J.
    Zohren, Stefan
    Roberts, Stephen J.
    ENTROPY, 2023, 25 (06)
  • [35] Hierarchical Correlations Replay for Continual Learning
    Wang, Qiang
    Liu, Jiayi
    Ji, Zhong
    Pang, Yanwei
    Zhang, Zhongfei
    KNOWLEDGE-BASED SYSTEMS, 2022, 250
  • [36] Memory Enhanced Replay for Continual Learning
    Xu, Guixun
    Guo, Wenhui
    Wang, Yanjiang
    2022 16TH IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP2022), VOL 1, 2022, : 218 - 222
  • [37] Continual compression model for online continual learning
    Ye, Fei
    Bors, Adrian G.
    APPLIED SOFT COMPUTING, 2024, 167
  • [38] Drinking From a Firehose: Continual Learning With Web-Scale Natural Language
    Hu, Hexiang
    Sener, Ozan
    Sha, Fei
    Koltun, Vladlen
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (05) : 5684 - 5696
  • [39] RETRACTED: Continual Learning Approach for Continuous Data Stream Analysis in Dynamic Environments (Retracted Article)
    Prasanna, K.
    Khan, Mudassir
    Alshahrani, Saeed M.
    Kiran, Ajmeera
    Reddy, P. Phanindra Kumar
    Alymani, Mofadal
    Babu, J. Chinna
    APPLIED SCIENCES-BASEL, 2023, 13 (14):
  • [40] Logarithmic Continual Learning
    Masarczyk, Wojciech
    Wawrzynski, Pawel
    Marczak, Daniel
    Deja, Kamil
    Trzcinski, Tomasz
    IEEE ACCESS, 2022, 10 : 117001 - 117010