Single-Head Lifelong Learning Based on Distilling Knowledge

被引:4
|
作者
Wang, Yen-Hsiang [1 ]
Lin, Chih-Yang [2 ]
Thaipisutikul, Tipajin [3 ]
Shih, Timothy K. [1 ]
机构
[1] Natl Cent Univ, Dept Comp Sci & Informat Engn, Taoyuan 320, Taiwan
[2] Yuan Ze Univ, Dept Elect Engn, Taoyuan 32003, Taiwan
[3] Mahidol Univ, Fac Informat & Commun Technol, Salaya 73170, Thailand
来源
IEEE ACCESS | 2022年 / 10卷
关键词
Task analysis; Neural networks; Training; Knowledge engineering; Data models; Testing; Predictive models; Lifelong learning; continuous learning; incremental learning; knowledge distillation; IMBALANCED DATA;
D O I
10.1109/ACCESS.2022.3155451
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Within the machine learning field, the main purpose of lifelong learning, also known as continuous learning, is to enable neural networks to learn continuously, as humans do. Lifelong learning accumulates the knowledge learned from previous tasks and transfers it to support the neural network in future tasks. This technique not only avoids the catastrophic forgetting problem with previous tasks when training new tasks, but also makes the model more robust with the temporal evolution. Motivated by the recent intervention of the lifelong learning technique, this paper presents a novel feature-based knowledge distillation method that differs from the existing methods of knowledge distillation in lifelong learning. Specifically, our proposed method utilizes the features from intermediate layers and compresses them in a unique way that involves global average pooling and fully connected layers. The authors then use the output of this branch network to deliver information from previous tasks to the model in the future. Extensive experiments show that our proposed model consistency outperforms the state-of-the-art baselines with the accuracy metric by at least two percent improvement under different experimental settings.
引用
收藏
页码:35469 / 35478
页数:10
相关论文
共 50 条
  • [31] Actual demands based knowledge intensive technological model for lifelong learning development in Vidzeme Region
    Osis, Kaspars
    Cakula, Sarma
    Kapenieks, Atis
    INTERNATIONAL JOURNAL OF EDUCATION AND INFORMATION TECHNOLOGIES, 2016, 10 : 25 - 30
  • [32] Meta-Learning Based Tasks Similarity Representation for Cross Domain Lifelong Learning
    Shen, Mingge
    Chen, Dehu
    Ren, Teng
    IEEE ACCESS, 2023, 11 : 36692 - 36701
  • [33] Patch-based Knowledge Distillation for Lifelong Person Re-Identification
    Sun, Zhicheng
    Mu, Yadong
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022,
  • [34] Knowledge Distillation-Based Representation Learning for Short-Utterance Spoken Language Identification
    Shen, Peng
    Lu, Xugang
    Li, Sheng
    Kawai, Hisashi
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 (28) : 2674 - 2683
  • [35] Lifelong Adaptive Machine Learning for Sensor-Based Human Activity Recognition Using Prototypical Networks
    Adaimi, Rebecca
    Thomaz, Edison
    SENSORS, 2022, 22 (18)
  • [36] A novel lifelong learning model based on cross domain knowledge extraction and transfer to classify underwater images
    Irfan, Muhammad
    Zheng Jiangbin
    Iqbal, Muhammad
    Arif, Muhammad Hassan
    INFORMATION SCIENCES, 2021, 552 : 80 - 101
  • [37] CoDERT: Distilling Encoder Representations with Co-learning for Transducer-based Speech Recognition
    Swaminathan, Rupak Vignesh
    King, Brian
    Strimel, Grant P.
    Droppo, Jasha
    Mouchtaris, Athanasios
    INTERSPEECH 2021, 2021, : 4543 - 4547
  • [38] Uncertainty-based modulation for lifelong learning
    Brna, Andrew P.
    Brown, Ryan C.
    Connolly, Patrick M.
    Simons, Stephen B.
    Shimizu, Renee E.
    Aguilar-Simon, Mario
    NEURAL NETWORKS, 2019, 120 : 129 - 142
  • [39] Sense-making: Information Literacy for Lifelong Learning and Knowledge Management
    Singh, Jagtar
    DESIDOC JOURNAL OF LIBRARY & INFORMATION TECHNOLOGY, 2008, 28 (02): : 13 - 17
  • [40] An Embedding-Based Approach to Rule Learning in Knowledge Graphs
    Omran, Pouya Ghiasnezhad
    Wang, Kewen
    Wang, Zhe
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2021, 33 (04) : 1348 - 1359