Lightweight image classifier using dilated and depthwise separable convolutions

被引:23
作者
Sun, Wei [1 ,2 ]
Zhang, Xiaorui [2 ,3 ]
He, Xiaozheng [4 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Automat, Nanjing 210044, Peoples R China
[2] Jiangsu Collaborat Innovat Ctr Atmospher Environm, Nanjing 210044, Peoples R China
[3] Jiangsu Engn Ctr Network Monitoring, Nanjing 210044, Peoples R China
[4] Rensselaer Polytech Inst, Dept Civil & Environm Engn, Troy, NY 12180 USA
来源
JOURNAL OF CLOUD COMPUTING-ADVANCES SYSTEMS AND APPLICATIONS | 2020年 / 9卷 / 01期
关键词
Classification accuracy; Cloud computing; Depthwise separable convolution; Dilated convolution; Lightweight neural network; IMPACTS; RECOMMENDATION; CHOICE; POLICY;
D O I
10.1186/s13677-020-00203-9
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The image classification based on cloud computing suffers from difficult deployment as the network depth and data volume increase. Due to the depth of the model and the convolution process of each layer will produce a great amount of calculation, the GPU and storage performance of the device are extremely demanding, and the GPU and storage devices equipped on the embedded and mobile terminals cannot support large models. So it is necessary to compress the model so that the model can be deployed on these devices. Meanwhile, traditional compression based methods often miss many global features during the compression process, resulting in low classification accuracy. To solve the problem, this paper proposes a lightweight neural network model based on dilated convolution and depthwise separable convolution with twenty-nine layers for image classification. The proposed model employs the dilated convolution to expand the receptive field during the convolution process while maintaining the number of convolution parameters, which can extract more high-level global semantic features to improve the classification accuracy. Also, the depthwise separable convolution is applied to reduce the network parameters and computational complexity in convolution operations, which reduces the size of the network. The proposed model introduces three hyperparameters: width multiplier, image resolution, and dilated rate, to compress the network on the premise of ensuring accuracy. The experimental results show that compared with GoogleNet, the network proposed in this paper improves the classification accuracy by nearly 1%, and the number of parameters is reduced by 3.7 million.
引用
收藏
页数:12
相关论文
共 45 条
[11]   Personal and societal impacts of motorcycle ban policy on motorcyclists' home-to-work morning commute in China [J].
Guo, Yuntao ;
Wang, Jian ;
Peeta, Srinivas ;
Anastasopoulos, Panagiotis Ch .
TRAVEL BEHAVIOUR AND SOCIETY, 2020, 19 :137-150
[12]   Impacts of internal migration, household registration system, and family planning policy on travel mode choice in China [J].
Guo, Yuntao ;
Wang, Jian ;
Peeta, Srinivas ;
Anastasopoulos, Panagiotis Ch. .
TRAVEL BEHAVIOUR AND SOCIETY, 2018, 13 :128-143
[13]  
Han S., 2016, INT C LEARN REPR ICL
[14]  
Han S, INT C LEARNING REPRE, V1, P14
[15]  
Han S, 2015, ADV NEUR IN, V28
[16]   Identity Mappings in Deep Residual Networks [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 :630-645
[17]   The nonaxismmetric endwall aerodynamic optimization design for a large turbine cascade with a midgap [J].
He, Lei ;
Liu, Hao ;
Zhu, Xiaocheng ;
Hua, Ouyang ;
Du, Zhaohui .
PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART G-JOURNAL OF AEROSPACE ENGINEERING, 2019, 233 (03) :1000-1010
[18]  
Howard A, 2017, INT C LEARN REPR, P1
[19]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]
[20]   Keynote: Small Neural Nets Are Beautiful: Enabling Embedded Systems with Small Deep-Neural-Network Architectures [J].
Iandola, Forrest ;
Keutzer, Kurt .
2017 INTERNATIONAL CONFERENCE ON HARDWARE/SOFTWARE CODESIGN AND SYSTEM SYNTHESIS (CODES+ISSS), 2017,