Accretionary Learning With Deep Neural Networks With Applications

被引:1
作者
Wei, Xinyu [1 ]
Juang, Biing-Hwang [1 ]
Wang, Ouya [2 ]
Zhou, Shenglong [3 ]
Li, Geoffrey Ye [2 ]
机构
[1] Georgia Inst Technol, Sch Elect & Comp Engn, Atlanta, GA 30332 USA
[2] Imperial Coll London, Dept Elect & Elect Engn, London SW7 2BX, England
[3] Beijing Jiaotong Univ, Sch Math & Stat, Beijing, Peoples R China
关键词
Artificial neural networks; Data models; Knowledge engineering; Task analysis; Training; Speech recognition; Learning systems; Deep learning; accretion learning; deep neural networks; pattern recognition; wireless communications; CLASSIFICATION;
D O I
10.1109/TCCN.2023.3342454
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
One of the fundamental limitations of Deep Neural Networks (DNN) is their inability to acquire and accumulate new cognitive capabilities in an incremental or progressive manner. When data appear from object classes not among the learned ones, a conventional DNN would not be able to recognize them due to the fundamental formulation that it assumes. A typical solution is to re-design and re-learn a new network, most likely an expanded one, for the expanded set of object classes. This process is quite different from that of a human learner. In this paper, we propose a new learning method named Accretionary Learning (AL) to emulate human learning, in that the set of object classes to be recognized need not be fixed, meaning it can grow as the situation arises without requiring an entire redesign of the system. The proposed learning structure is modularized, and can dynamically expand to learn and register new knowledge, as the set of objects grows in size. AL does not forget previous knowledge when learning new data classes. We show that the structure and its learning methodology lead to a system that can grow to cope with increased cognitive complexity while providing stable and superior overall performance.
引用
收藏
页码:660 / 673
页数:14
相关论文
共 50 条
[41]   Breast Cancer Prognosis Based on Transfer Learning Techniques in Deep Neural Networks [J].
Diwakaran, M. ;
Surendran, D. .
INFORMATION TECHNOLOGY AND CONTROL, 2023, 52 (02) :381-396
[42]   Backdoor Attacks to Deep Neural Networks: A Survey of the Literature, Challenges, and Future Research Directions [J].
Mengara, Orson ;
Avila, Anderson ;
Falk, Tiago H. .
IEEE ACCESS, 2024, 12 :29004-29023
[43]   Understanding Error Propagation in Deep-Learning Neural Networks' Accelerators and Applications [J].
Li, Guanpeng ;
Hari, Siva Kumar Sastry ;
Sullivan, Michael B. ;
Tsai, Timothy ;
Pattabiraman, Karthik ;
Emer, Joel S. ;
Keckler, Stephen W. .
IEEE DESIGN & TEST, 2025, 42 (03) :7-13
[44]   Classification of Vernacular Landscape Elements and Design Applications Based on Deep Learning Fully Convolutional Neural Networks [J].
Li, Ya ;
Zhang, Yun .
ADVANCES IN CIVIL ENGINEERING, 2025, 2025 (01)
[45]   Deep Learning with Random Neural Networks [J].
Gelenbe, Erol ;
Yin, Yongha .
PROCEEDINGS OF SAI INTELLIGENT SYSTEMS CONFERENCE (INTELLISYS) 2016, VOL 2, 2018, 16 :450-462
[46]   Deep learning in spiking neural networks [J].
Tavanaei, Amirhossein ;
Ghodrati, Masoud ;
Kheradpisheh, Saeed Reza ;
Masquelier, Timothee ;
Maida, Anthony .
NEURAL NETWORKS, 2019, 111 :47-63
[47]   Fast learning in Deep Neural Networks [J].
Chandra, B. ;
Sharma, Rajesh K. .
NEUROCOMPUTING, 2016, 171 :1205-1215
[48]   Explaining Deep Neural Networks and Beyond: A Review of Methods and Applications [J].
Samek, Wojciech ;
Montavon, Gregoire ;
Lapuschkin, Sebastian ;
Anders, Christopher J. ;
Mueller, Klaus-Robert .
PROCEEDINGS OF THE IEEE, 2021, 109 (03) :247-278
[49]   Hierarchical Training of Deep Neural Networks Using Early Exiting [J].
Sepehri, Yamin ;
Pad, Pedram ;
Yuzuguler, Ahmet Caner ;
Frossard, Pascal ;
Dunbar, L. Andrea .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (04) :6271-6285
[50]   A Gradient Boosting Approach for Training Convolutional and Deep Neural Networks [J].
Emami, Seyedsaman ;
Martinez-Munoz, Gonzalo .
IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2023, 4 :313-321