A Study on Catastrophic Forgetting in Deep LSTM Networks

被引:17
|
作者
Schak, Monika [1 ]
Gepperth, Alexander [1 ]
机构
[1] Univ Appl Sci Fulda, D-36037 Fulda, Germany
来源
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: DEEP LEARNING, PT II | 2019年 / 11728卷
关键词
LSTM; Catastrophic Forgetting;
D O I
10.1007/978-3-030-30484-3_56
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a systematic study of Catastrophic Forgetting (CF), i.e., the abrupt loss of previously acquired knowledge, when retraining deep recurrent LSTM networks with new samples. CF has recently received renewed attention in the case of feed-forward DNNs, and this article is the first work that aims to rigorously establish whether deep LSTM networks are afflicted by CF as well, and to what degree. In order to test this fully, training is conducted using a wide variety of high-dimensional image-based sequence classification tasks derived from established visual classification benchmarks (MNIST, Devanagari, FashionMNIST and EMNIST). We find that the CF effect occurs universally, without exception, for deep LSTM-based sequence classifiers, regardless of the construction and provenance of sequences. This leads us to conclude that LSTMs, just like DNNs, are fully affected by CF, and that further research work needs to be conducted in order to determine how to avoid this effect (which is not a goal of this study).
引用
收藏
页码:714 / 728
页数:15
相关论文
共 50 条
  • [41] EXPLOITING LSTM STRUCTURE IN DEEP NEURAL NETWORKS FOR SPEECH RECOGNITION
    He, Tianxing
    Droppo, Jasha
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 5445 - 5449
  • [42] LSTM Deep Neural Networks Postfiltering for Enhancing Synthetic Voices
    Coto-Jimenez, Marvin
    Goddard-Close, John
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2018, 32 (01)
  • [43] Neural networks with a self-refreshing memory: knowledge transfer in sequential learning tasks without catastrophic forgetting
    Ans, B
    Rousset, S
    CONNECTION SCIENCE, 2000, 12 (01) : 1 - 19
  • [44] Deep image captioning using an ensemble of CNN and LSTM based deep neural networks
    Alzubi, Jafar A.
    Jain, Rachna
    Nagrath, Preeti
    Satapathy, Suresh
    Taneja, Soham
    Gupta, Paras
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 40 (04) : 5761 - 5769
  • [45] Relevant Question Answering in Community Based Networks Using Deep LSTM Neural Networks
    Karimi, Elaheh
    Majidi, Babak
    Manzuri, Mohammad Taghi
    2019 7TH IRANIAN JOINT CONGRESS ON FUZZY AND INTELLIGENT SYSTEMS (CFIS), 2019, : 36 - 40
  • [46] Assessment of catastrophic forgetting in continual credit card fraud detection
    Lebichot, B.
    Siblini, W.
    Paldino, G. M.
    Le Borgne, Y. -A.
    Oble, F.
    Bontempi, G.
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 249
  • [47] Soft sensor development and applications based on LSTM in deep neural networks
    Ke, Wensi
    Huang, Dexian
    Yang, Fan
    Jiang, Yongheng
    2017 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2017, : 3468 - 3473
  • [48] LSTM Deep Neural Networks Postfiltering for Improving the Quality of Synthetic Voices
    Coto-Jimenez, Marvin
    Goddard-Close, John
    PATTERN RECOGNITION (MCPR 2016), 2016, 9703 : 280 - 289
  • [49] Overcoming Catastrophic Forgetting in Continual Learning by Exploring Eigenvalues of Hessian Matrix
    Kong, Yajing
    Liu, Liu
    Chen, Huanhuan
    Kacprzyk, Janusz
    Tao, Dacheng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 16196 - 16210
  • [50] Detecting Changes and Avoiding Catastrophic Forgetting in Dynamic Partially Observable Environments
    Dick, Jeffery
    Ladosz, Pawel
    Ben-Iwhiwhu, Eseoghene
    Shimadzu, Hideyasu
    Kinnell, Peter
    Pilly, Praveen K.
    Kolouri, Soheil
    Soltoggio, Andrea
    FRONTIERS IN NEUROROBOTICS, 2020, 14