Learning Automata Based Incremental Learning Method for Deep Neural Networks

被引:15
作者
Guo, Haonan [1 ]
Wang, Shilin [2 ,3 ,4 ]
Fan, Jianxun [5 ]
Li, Shenghong [2 ,3 ,4 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Dept Elect Engn, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Sch Cyber Secur, Shanghai 200240, Peoples R China
[3] Shanghai Jiao Tong Univ, Sch Elect Informat & Elect Engn, Shanghai 200240, Peoples R China
[4] Shanghai Jiao Tong Univ, Artificial Intelligence Inst, Shanghai 200240, Peoples R China
[5] Beijing Univ Posts & Telecommun, Elect Sci & Technol Dept, Beijing 100086, Peoples R China
关键词
Supervised learning; incremental learning; learning automata;
D O I
10.1109/ACCESS.2019.2907645
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning methods have got fantastic performance on lots of large-scale datasets for machine learning tasks, such as visual recognition and neural language processing. Most of the progress on deep learning in recent years lied on supervised learning, for which the whole dataset with respect to a specific task should be well-prepared before training. However, in the real-world scenario, the labeled data associated with the assigned classes are always gathered incrementally over time, since it is cumbersome work to collect and annotate the training data manually. This suggests the manner of sequentially training on a series of datasets with gradually added training samples belonging to new classes, which is called incremental learning. In this paper, we proposed an effective incremental training method based on learning automata for deep neural networks. The main thought is to train a deep model with dynamic connections which can be either "activated'' or "deactivated'' on different datasets of the incremental training stages. Our proposed method can relieve the destruction of old features while learning new features for the newly added training samples, which can lead to better training performance on the incremental learning stage. The experiments on MNIST and CIFAR-100 demonstrated that our method can be implemented for deep neural models in a long sequence of incremental training stages and can achieve superior performance than training from scratch and the fine-tuning method.
引用
收藏
页码:41164 / 41171
页数:8
相关论文
共 42 条
[1]  
[Anonymous], P 3 INT C LEARNING R
[2]  
[Anonymous], P IEEE C COMP VIS PA
[3]  
[Anonymous], PROC CVPR IEEE
[4]  
[Anonymous], 2015, Nature, DOI [10.1038/nature14539, DOI 10.1038/NATURE14539]
[5]  
[Anonymous], 2017, COMMUN ACM, DOI DOI 10.1145/3065386
[6]  
[Anonymous], PROC CVPR IEEE
[7]  
[Anonymous], IEEE ACM T AUDIO SPE
[8]  
[Anonymous], 2015, ARXIV PREPRINT ARXIV
[9]  
[Anonymous], 2014, P ADV NEUR INF PROC
[10]  
[Anonymous], 2013, CoRR