From MNIST to ImageNet and back: benchmarking continual curriculum learning

被引:3
|
作者
Faber, Kamil [1 ]
Zurek, Dominik [1 ]
Pietron, Marcin [1 ]
Japkowicz, Nathalie [3 ]
Vergari, Antonio [2 ]
Corizzo, Roberto [3 ]
机构
[1] AGH Univ Krakow, PL-30059 Krakow, Poland
[2] Univ Edinburgh, Edinburgh EH8 9AB, Scotland
[3] Amer Univ, Washington, DC 20016 USA
关键词
Continual learning; Lifelong learning; Curriculum learning; Neural networks; Computer vision; Image classification;
D O I
10.1007/s10994-024-06524-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continual learning (CL) is one of the most promising trends in recent machine learning research. Its goal is to go beyond classical assumptions in machine learning and develop models and learning strategies that present high robustness in dynamic environments. This goal is realized by designing strategies that simultaneously foster the incorporation of new knowledge while avoiding forgetting past knowledge. The landscape of CL research is fragmented into several learning evaluation protocols, comprising different learning tasks, datasets, and evaluation metrics. Additionally, the benchmarks adopted so far are still distant from the complexity of real-world scenarios, and are usually tailored to highlight capabilities specific to certain strategies. In such a landscape, it is hard to clearly and objectively assess models and strategies. In this work, we fill this gap for CL on image data by introducing two novel CL benchmarks that involve multiple heterogeneous tasks from six image datasets, with varying levels of complexity and quality. Our aim is to fairly evaluate current state-of-the-art CL strategies on a common ground that is closer to complex real-world scenarios. We additionally structure our benchmarks so that tasks are presented in increasing and decreasing order of complexity-according to a curriculum-in order to evaluate if current CL models are able to exploit structure across tasks. We devote particular emphasis to providing the CL community with a rigorous and reproducible evaluation protocol for measuring the ability of a model to generalize and not to forget while learning. Furthermore, we provide an extensive experimental evaluation showing that popular CL strategies, when challenged with our proposed benchmarks, yield sub-par performance, high levels of forgetting, and present a limited ability to effectively leverage curriculum task ordering. We believe that these results highlight the need for rigorous comparisons in future CL works as well as pave the way to design new CL strategies that are able to deal with more complex scenarios.
引用
收藏
页码:8137 / 8164
页数:28
相关论文
共 43 条
  • [1] Reinforcement learning for quadrupedal locomotion with design of continual-hierarchical curriculum
    Kobayashi, Taisuke
    Sugino, Toshiki
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2020, 95 (95)
  • [2] ImageNet-Patch: A dataset for benchmarking machine learning robustness against adversarial patches
    Pintor, Maura
    Angioni, Daniele
    Sotgiu, Angelo
    Demetrio, Luca
    Demontis, Ambra
    Biggio, Battista
    Roli, Fabio
    PATTERN RECOGNITION, 2023, 134
  • [3] Continual Learning From a Stream of APIs
    Yang, Enneng
    Wang, Zhenyi
    Shen, Li
    Yin, Nan
    Liu, Tongliang
    Guo, Guibing
    Wang, Xingwei
    Tao, Dacheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 11432 - 11445
  • [4] Continual Learning-Based MIMO Channel Estimation: A Benchmarking Study
    Akrout, Mohamed
    Feriani, Amal
    Bellili, Faouzi
    Mezghani, Amine
    Hossain, Ekram
    ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 2631 - 2636
  • [5] Drinking From a Firehose: Continual Learning With Web-Scale Natural Language
    Hu, Hexiang
    Sener, Ozan
    Sha, Fei
    Koltun, Vladlen
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (05) : 5684 - 5696
  • [6] Continual learning from demonstration of robotics skills
    Auddy, Sayantan
    Hollenstein, Jakob
    Saveriano, Matteo
    Rodriguez-Sanchez, Antonio
    Piater, Justus
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2023, 165
  • [7] Continual Learning for Fake News Detection from Social Media
    Han, Yi
    Karunasekera, Shanika
    Leckie, Christopher
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT II, 2021, 12892 : 372 - 384
  • [8] Tensor decision trees for continual learning from drifting data streams
    Bartosz Krawczyk
    Machine Learning, 2021, 110 : 3015 - 3035
  • [9] Tensor Decision Trees for Continual Learning from Drifting Data Streams
    Krawczyk, Bartosz
    2021 IEEE 8TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2021,
  • [10] Tensor decision trees for continual learning from drifting data streams
    Krawczyk, Bartosz
    MACHINE LEARNING, 2021, 110 (11-12) : 3015 - 3035