Learning dynamics of kernel-based deep neural networks in manifolds

被引:0
作者
Wu, Wei [1 ,3 ]
Jing, Xiaoyuan [1 ,2 ]
Du, Wencai [4 ]
Chen, Guoliang [5 ]
机构
[1] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
[2] Guangdong Univ Petrochem Technol, Sch Comp, Maoming 525000, Peoples R China
[3] Chinese Acad Sci, Inst Deep Sea Sci & Engn, Sanya 572000, Peoples R China
[4] City Univ Macau, Inst Data Sci, Macau 999078, Peoples R China
[5] Shenzhen Univ, Coll Comp Sci & Software Engn, Shenzhen 518060, Peoples R China
基金
中国国家自然科学基金;
关键词
learning dynamics; kernel-based convolution; manifolds; control model; network stability; SINGULARITIES; WORKS;
D O I
10.1007/s11432-020-3022-3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional neural networks (CNNs) obtain promising results via layered kernel convolution and pooling operations, yet the learning dynamics of the kernel remain obscure. We propose a continuous form to describe kernel-based convolutions through integration in neural manifolds. The status of spatial expression is proposed to analyze the stability of kernel-based CNNs. We divide CNN dynamics into the three stages of unstable vibration, collaborative adjusting, and stabilized fluctuation. According to the system control matrix of the kernel, the kernel-based CNN training proceeds via the unstable and stable status and is verified by numerical experiments.
引用
收藏
页数:15
相关论文
共 26 条
[1]   Natural gradient works efficiently in learning [J].
Amari, S .
NEURAL COMPUTATION, 1998, 10 (02) :251-276
[2]   Singularities affect dynamics of learning in neuromanifolds [J].
Amari, Shun-ichi ;
Park, Hyeyoung ;
Ozeki, Tomoko .
NEURAL COMPUTATION, 2006, 18 (05) :1007-1065
[3]   Why Deep Learning Works: A Manifold Disentanglement Perspective [J].
Brahma, Pratik Prabhanjan ;
Wu, Dapeng ;
She, Yiyuan .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (10) :1997-2008
[4]   The minicolumn hypothesis in neuroscience [J].
Buxhoeveden, DP ;
Casanova, MF .
BRAIN, 2002, 125 :935-951
[5]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[6]   Dynamics of learning in multilayer perceptrons near singularities [J].
Cousseau, Florent ;
Ozeki, Tomoko ;
Amari, Shun-ichi .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (08) :1313-1328
[7]   A Novel Criterion for Global Asymptotic Stability of Neutral-Type Neural Networks with Discrete Time Delays [J].
Faydasicok, Ozlem ;
Arik, Sabri .
NEURAL INFORMATION PROCESSING (ICONIP 2018), PT II, 2018, 11302 :353-360
[8]  
Guo WL, 2018, J MACH LEARN RES, V19
[9]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[10]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90