Single-Head Lifelong Learning Based on Distilling Knowledge

被引:4
|
作者
Wang, Yen-Hsiang [1 ]
Lin, Chih-Yang [2 ]
Thaipisutikul, Tipajin [3 ]
Shih, Timothy K. [1 ]
机构
[1] Natl Cent Univ, Dept Comp Sci & Informat Engn, Taoyuan 320, Taiwan
[2] Yuan Ze Univ, Dept Elect Engn, Taoyuan 32003, Taiwan
[3] Mahidol Univ, Fac Informat & Commun Technol, Salaya 73170, Thailand
来源
IEEE ACCESS | 2022年 / 10卷
关键词
Task analysis; Neural networks; Training; Knowledge engineering; Data models; Testing; Predictive models; Lifelong learning; continuous learning; incremental learning; knowledge distillation; IMBALANCED DATA;
D O I
10.1109/ACCESS.2022.3155451
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Within the machine learning field, the main purpose of lifelong learning, also known as continuous learning, is to enable neural networks to learn continuously, as humans do. Lifelong learning accumulates the knowledge learned from previous tasks and transfers it to support the neural network in future tasks. This technique not only avoids the catastrophic forgetting problem with previous tasks when training new tasks, but also makes the model more robust with the temporal evolution. Motivated by the recent intervention of the lifelong learning technique, this paper presents a novel feature-based knowledge distillation method that differs from the existing methods of knowledge distillation in lifelong learning. Specifically, our proposed method utilizes the features from intermediate layers and compresses them in a unique way that involves global average pooling and fully connected layers. The authors then use the output of this branch network to deliver information from previous tasks to the model in the future. Extensive experiments show that our proposed model consistency outperforms the state-of-the-art baselines with the accuracy metric by at least two percent improvement under different experimental settings.
引用
收藏
页码:35469 / 35478
页数:10
相关论文
共 50 条
  • [41] Comprehensive learning and adaptive teaching: Distilling multi-modal knowledge for pathological glioma grading
    Xing, Xiaohan
    Zhu, Meilu
    Chen, Zhen
    Yuan, Yixuan
    MEDICAL IMAGE ANALYSIS, 2024, 91
  • [42] MicroBERT: Distilling MoE-Based Knowledge from BERT into a Lighter Model
    Zheng, Dashun
    Li, Jiaxuan
    Yang, Yunchu
    Wang, Yapeng
    Pang, Patrick Cheong-Iao
    APPLIED SCIENCES-BASEL, 2024, 14 (14):
  • [43] Knowledge Tracing Through Enhanced Questions and Directed Learning Interaction Based on Multigraph Embeddings in Intelligent Tutoring Systems
    Qiu, Liqing
    Wang, Lulu
    IEEE TRANSACTIONS ON EDUCATION, 2025, 68 (01) : 43 - 56
  • [44] Global Information-Based Lifelong Dictionary Learning for Multimode Process Monitoring
    Chen, Zixuan
    Huang, Keke
    Wu, Dehao
    Yang, Chunhua
    Gui, Weihua
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2024, 54 (12): : 7182 - 7194
  • [45] Learning face super-resolution through identity features and distilling facial prior knowledge
    Tomara, Anurag Singh
    Arya, K. V.
    Rajput, Shyam Singh
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 262
  • [46] Education of the twenty-first century: New professionalism in lifelong learning, knowledge development and knowledge sharing
    van Weert, Tom J.
    EDUCATION AND INFORMATION TECHNOLOGIES, 2006, 11 (3-4) : 217 - 237
  • [47] Education of the twenty-first century: New professionalism in lifelong learning, knowledge development and knowledge sharing
    Tom J. van Weert
    Education and Information Technologies, 2006, 11 (3-4) : 217 - 237
  • [48] Local knowledge, global knowledge: The role of local language literacy for lifelong learning in rural African contexts
    Trudell, Joel
    Cheffy, Ian
    INTERNATIONAL REVIEW OF EDUCATION, 2019, 65 (03) : 409 - 425
  • [49] Local knowledge, global knowledge: The role of local language literacy for lifelong learning in rural African contexts
    Joel Trudell
    Ian Cheffy
    International Review of Education, 2019, 65 : 409 - 425
  • [50] A Cooking Knowledge Graph and Benchmark for Question Answering Evaluation in Lifelong Learning Scenarios
    Veron, Mathilde
    Penas, Anselmo
    Echegoyen, Guillermo
    Banerjee, Somnath
    Ghannay, Sahar
    Rosset, Sophie
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2020), 2020, 12089 : 94 - 101