Small and Slim Deep Convolutional Neural Network for Mobile Device

被引:28
作者
Winoto, Amadeus Suryo [1 ]
Kristianus, Michael [1 ]
Premachandra, Chinthaka [2 ]
机构
[1] BINUS Univ, Dept Comp Sci, Jakarta 11480, Indonesia
[2] Shibaura Inst Technol, Grad Sch Engn & Sci, Dept Elect Engn, Sch Engn, Tokyo 1358548, Japan
关键词
Artificial neural network; image recognition; machine learning; deep learning;
D O I
10.1109/ACCESS.2020.3005161
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent development of deep convolutional neural networks (DCNN) devoted in creating a slim model for devices with lower specification such as embedded, mobile hardware, or microcomputer. Slim model can be achieved by minimizing computational complexity which theoretically will make processing time faster. Therefore, our focus is to build an architecture with minimum floating-point operation per second (FLOPs). In this work, we propose a small and slim architecture which later will be compared to state-of-the-art models. This architecture will be implemented into two models which are CustomNet and CustomNet2. Each of these models implements 3 convolutional blocks which reduce the computational complexity while maintains its accuracy and able to compete with state-of-the-art DCNN models. These models will be trained using ImageNet, CIFAR 10, CIFAR 100 and other datasets. The result will be compared based on accuracy, complexity, size, processing time, and trainable parameter. From the result, we found that one of our models which is CustomNet2, is better than MobileNet, MobileNet-v2, DenseNet, NASNetMobile in accuracy, trainable parameter, and complexity. For future implementation, this architecture can be adapted using region based DCNN for multiple object detection.
引用
收藏
页码:125210 / 125222
页数:13
相关论文
共 41 条
[31]   MobileNetV2: Inverted Residuals and Linear Bottlenecks [J].
Sandler, Mark ;
Howard, Andrew ;
Zhu, Menglong ;
Zhmoginov, Andrey ;
Chen, Liang-Chieh .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4510-4520
[32]  
Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
[33]  
Stallkamp J, 2011, 2011 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), P1453, DOI 10.1109/IJCNN.2011.6033395
[34]  
Subhi MA, 2018, IEEE EMBS CONF BIO, P284, DOI 10.1109/IECBES.2018.8626720
[35]  
Sun S., 2019, ARXIV190103495
[36]  
Szegedy C, 2014, Arxiv, DOI [arXiv:1312.6199, DOI 10.1109/CVPR.2015.7298594]
[37]   Rethinking the Inception Architecture for Computer Vision [J].
Szegedy, Christian ;
Vanhoucke, Vincent ;
Ioffe, Sergey ;
Shlens, Jon ;
Wojna, Zbigniew .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2818-2826
[38]   Victim Detection Using UAV with On-board Voice Recognition System [J].
Yamazaki, Yuki ;
Tamaki, Masaya ;
Premachandra, C. ;
Perera, C. J. ;
Sumathipala, S. ;
Sudantha, B. H. .
2019 THIRD IEEE INTERNATIONAL CONFERENCE ON ROBOTIC COMPUTING (IRC 2019), 2019, :555-559
[39]  
Zhang H, 2019, 2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), P122, DOI [10.1109/aicas.2019.8771540, 10.1109/AICAS.2019.8771540]
[40]  
Zhou Z.-H., 2012, ENSEMBLE METHODS FDN, DOI [10.1201/b12207, DOI 10.1201/B12207]