SRNET: A Shallow Skip Connection Based Convolutional Neural Network Design for Resolving Singularities

被引:0
作者
Robail Yasrab
机构
[1] University of Nottingham,Computer Vision Laboratory, School of Computer Science
来源
Journal of Computer Science and Technology | 2019年 / 34卷
关键词
convolutional neural network (CNN); wide residual network (WRN); dropout; skip connection; deep neural network (DNN);
D O I
暂无
中图分类号
学科分类号
摘要
Convolutional neural networks (CNNs) have shown tremendous progress and performance in recent years. Since emergence, CNNs have exhibited excellent performance in most of classification and segmentation tasks. Currently, the CNN family includes various architectures that dominate major vision-based recognition tasks. However, building a neural network (NN) by simply stacking convolution blocks inevitably limits its optimization ability and introduces overfitting and vanishing gradient problems. One of the key reasons for the aforementioned issues is network singularities, which have lately caused degenerating manifolds in the loss landscape. This situation leads to a slow learning process and lower performance. In this scenario, the skip connections turned out to be an essential unit of the CNN design to mitigate network singularities. The proposed idea of this research is to introduce skip connections in NN architecture to augment the information flow, mitigate singularities and improve performance. This research experimented with different levels of skip connections and proposed the placement strategy of these links for any CNN. To prove the proposed hypothesis, we designed an experimental CNN architecture, named as Shallow Wide ResNet or SRNet, as it uses wide residual network as a base network design. We have performed numerous experiments to assess the validity of the proposed idea. CIFAR-10 and CIFAR-100, two well-known datasets are used for training and testing CNNs. The final empirical results have shown a great many of promising outcomes in terms of performance, efficiency and reduction in network singularities issues.
引用
收藏
页码:924 / 938
页数:14
相关论文
共 40 条
  • [1] Russakovsky O(2015)ImageNet large scale visual recognition challenge International Journal of Computer Vision 115 211-252
  • [2] Deng J(2015)Deep learning Nature 521 436-444
  • [3] Su H(1992)Learning complex, extended sequences using the principle of history compression Neural Computation 4 234-242
  • [4] Krause J(2008)Dynamics of learning near singularities in layered networks Neural Computation 20 813-843
  • [5] Satheesh S(2006)Singularities affect dynamics of learning in neuromanifolds Neural Computation 18 1007-1065
  • [6] Ma S(2018)Residual networks of residual networks: Multilevel residual networks IEEE Transactions on Circuits and Systems for Video Technology 28 1303-1314
  • [7] Berg AC(1994)Learning long-term dependencies with gradient descent is difficult IEEE Transactions on Neural Networks 5 157-166
  • [8] LeCun Y(2014)Dropout: A simple way to prevent neural networks from overfitting Journal of Machine Learning Research 15 1929-1958
  • [9] Yoshua B(1989)Backpropagation applied to handwritten zip code recognition Neural Computation 1 541-551
  • [10] Geoffrey EH(undefined)undefined undefined undefined undefined-undefined