Lightweight image classifier using dilated and depthwise separable convolutions

被引:23
作者
Sun, Wei [1 ,2 ]
Zhang, Xiaorui [2 ,3 ]
He, Xiaozheng [4 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Automat, Nanjing 210044, Peoples R China
[2] Jiangsu Collaborat Innovat Ctr Atmospher Environm, Nanjing 210044, Peoples R China
[3] Jiangsu Engn Ctr Network Monitoring, Nanjing 210044, Peoples R China
[4] Rensselaer Polytech Inst, Dept Civil & Environm Engn, Troy, NY 12180 USA
来源
JOURNAL OF CLOUD COMPUTING-ADVANCES SYSTEMS AND APPLICATIONS | 2020年 / 9卷 / 01期
关键词
Classification accuracy; Cloud computing; Depthwise separable convolution; Dilated convolution; Lightweight neural network; IMPACTS; RECOMMENDATION; CHOICE; POLICY;
D O I
10.1186/s13677-020-00203-9
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The image classification based on cloud computing suffers from difficult deployment as the network depth and data volume increase. Due to the depth of the model and the convolution process of each layer will produce a great amount of calculation, the GPU and storage performance of the device are extremely demanding, and the GPU and storage devices equipped on the embedded and mobile terminals cannot support large models. So it is necessary to compress the model so that the model can be deployed on these devices. Meanwhile, traditional compression based methods often miss many global features during the compression process, resulting in low classification accuracy. To solve the problem, this paper proposes a lightweight neural network model based on dilated convolution and depthwise separable convolution with twenty-nine layers for image classification. The proposed model employs the dilated convolution to expand the receptive field during the convolution process while maintaining the number of convolution parameters, which can extract more high-level global semantic features to improve the classification accuracy. Also, the depthwise separable convolution is applied to reduce the network parameters and computational complexity in convolution operations, which reduces the size of the network. The proposed model introduces three hyperparameters: width multiplier, image resolution, and dilated rate, to compress the network on the premise of ensuring accuracy. The experimental results show that compared with GoogleNet, the network proposed in this paper improves the classification accuracy by nearly 1%, and the number of parameters is reduced by 3.7 million.
引用
收藏
页数:12
相关论文
共 45 条
[1]  
[Anonymous], 2016, CORR
[2]  
[Anonymous], 2016, ICLR
[3]   Bigtable: A distributed storage system for structured data [J].
Chang, Fay ;
Dean, Jeffrey ;
Ghemawat, Sanjay ;
Hsieh, Wilson C. ;
Wallach, Deborah A. ;
Burrows, Mike ;
Chandra, Tushar ;
Fikes, Andrew ;
Gruber, Robert E. .
ACM TRANSACTIONS ON COMPUTER SYSTEMS, 2008, 26 (02)
[4]   Energy Efficient Dynamic Offloading in Mobile Edge Computing for Internet of Things [J].
Chen, Ying ;
Zhang, Ning ;
Zhang, Yongchao ;
Chen, Xin ;
Wu, Wen ;
Shen, Xuemin .
IEEE TRANSACTIONS ON CLOUD COMPUTING, 2021, 9 (03) :1050-1060
[5]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[6]  
Dean J, 2004, USENIX ASSOCIATION PROCEEDINGS OF THE SIXTH SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION (OSDE '04), P137
[7]   Wireless Channel Parameter Estimation Algorithms: Recent Advances and Future Challenges [J].
Feng, Rui ;
Liu, Yu ;
Huang, Jie ;
Sun, Jian ;
Wang, Cheng-Xiang ;
Goussetis, George .
CHINA COMMUNICATIONS, 2018, 15 (05) :211-228
[8]  
Ghemawat S, 2003, SOSP, P29
[9]   Recent advances in convolutional neural networks [J].
Gu, Jiuxiang ;
Wang, Zhenhua ;
Kuen, Jason ;
Ma, Lianyang ;
Shahroudy, Amir ;
Shuai, Bing ;
Liu, Ting ;
Wang, Xingxing ;
Wang, Gang ;
Cai, Jianfei ;
Chen, Tsuhan .
PATTERN RECOGNITION, 2018, 77 :354-377
[10]   Impacts of personalized accessibility information on residential location choice and travel behavior [J].
Guo, Yuntao ;
Peeta, Srinivas .
TRAVEL BEHAVIOUR AND SOCIETY, 2020, 19 :99-111