Image classification based on RESNET

被引:49
作者
Liang, Jiazhi [1 ]
机构
[1] Northwest MuZu Univ, Comp Technol, Lanzhou, Gansu, Peoples R China
来源
2020 3RD INTERNATIONAL CONFERENCE ON COMPUTER INFORMATION SCIENCE AND APPLICATION TECHNOLOGY (CISAT) 2020 | 2020年 / 1634卷
关键词
D O I
10.1088/1742-6596/1634/1/012110
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
at present, neural networks are becoming more and more complex, from several layers to dozens of layers or even more than 100 layers.The main advantage of deep network is that it can express very complex functions. It can learn features from different levels of abstraction, such as edge features at lower levels and complex features at higher levels.However, the use of deep networks is not always effective, because there is a very big obstacle - the disappearance of gradients: in very deep networks, gradient signals tend to approach zero very quickly, which makes the gradient descent process extremely slow. Specifically, in the process of gradient descent, the weight matrix product operation must be carried out in every step of back propagation from the last layer to the first layer, so that the gradient will drop exponentially to 0.(in rare cases, there is the problem of gradient explosion, that is, the gradient grows exponentially to the overflow in the process of propagation). Therefore, in the process of training, it will be found that with the increase of the number of layers, the rate of gradient decrease increases.Therefore, by deepening the network, although it can express any complex function, but in fact, with the increase of network layers, we are more and more difficult to train the network, until the proposal of residual network, which makes it possible to train deeper network[1].
引用
收藏
页数:6
相关论文
共 9 条
[1]  
[Anonymous], 2017, ARXIV171200559
[2]  
[Anonymous], 2018, ARXIV180711164
[3]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[4]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/TPAMI.2019.2913372, 10.1109/CVPR.2018.00745]
[5]   CondenseNet: An Efficient DenseNet using Learned Group Convolutions [J].
Huang, Gao ;
Liu, Shichen ;
van der Maaten, Laurens ;
Weinberger, Kilian Q. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :2752-2761
[6]   MobileNetV2: Inverted Residuals and Linear Bottlenecks [J].
Sandler, Mark ;
Howard, Andrew ;
Zhu, Menglong ;
Zhmoginov, Andrey ;
Chen, Liang-Chieh .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4510-4520
[7]   Lung nodule classification using local kernel regression models with out-of-sample extension [J].
Wei, Guohui ;
Ma, He ;
Qian, Wei ;
Han, Fangfang ;
Jiang, Hongyang ;
Qi, Shouliang ;
Qiu, Min .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2018, 40 :1-9
[8]   ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices [J].
Zhang, Xiangyu ;
Zhou, Xinyu ;
Lin, Mengxiao ;
Sun, Ran .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :6848-6856
[9]   Learning Transferable Architectures for Scalable Image Recognition [J].
Zoph, Barret ;
Vasudevan, Vijay ;
Shlens, Jonathon ;
Le, Quoc V. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8697-8710