MOO-DNAS: Efficient Neural Network Design via Differentiable Architecture Search Based on Multi-Objective Optimization

被引:10
作者
Wei, Hui [1 ,2 ]
Lee, Feifei [1 ,2 ]
Hu, Chunyan [3 ]
Chen, Qiu [4 ]
机构
[1] Univ Shanghai Sci & Technol, Sch Med Instrument & Food Engn, Shanghai Engn Res Ctr Assist Devices, Shanghai 200093, Peoples R China
[2] Univ Shanghai Sci & Technol, Rehabil Engn & Technol Inst, Shanghai 200093, Peoples R China
[3] Univ Shanghai Sci & Technol, Sch Opt Elect & Comp Engineer, Shanghai 200093, Peoples R China
[4] Kogakuin Univ, Grad Sch Engn, Elect Engn & Elect, Tokyo 1638677, Japan
关键词
Computer architecture; Neural networks; Optimization; Computational modeling; Search problems; Convolutional neural networks; Complexity theory; Differentiable neural architecture search; CNNs; multi-objective optimization; accuracy-efficiency trade-off; PARTICLE SWARM OPTIMIZATION;
D O I
10.1109/ACCESS.2022.3148323
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The progress devoted to improving the performance of neural networks has come at a high price in terms of cost and experience. Fortunately, the emergence of Neural Architecture Search improves the speed of network design, but most excellent works only optimize for high accuracy without penalizing the model complexity. In this paper, we propose an efficient CNN architecture search framework, MOO-DNAS, with multi-objective optimization based on differentiable neural architecture search. The main goal is to trade off two competing objectives, classification accuracy and network latency, so that the search algorithm is able to discover an efficient model while maintaining high accuracy. In order to achieve a better implementation, we construct a novel factorized hierarchical search space to support layer variety and hardware friendliness. Furthermore, a robust sampling strategy named "hard-sampling" is proposed to obtain final structures with higher average performance by keeping the highest scoring operator. Experimental results on the benchmark datasets MINST, CIFAR10 and CIFAR100 demonstrate the effectiveness of the proposed method. The searched architectures, MOO-DNAS-Nets, achieve advanced accuracy with fewer parameters and FLOPs, and the search cost is less than one GPU-day.
引用
收藏
页码:14195 / 14207
页数:13
相关论文
共 58 条
[1]  
Baker B., 2017, INT C LEARNING REPRE
[2]  
Cai H, 2019, INT C LEARN REPR
[3]   Pedestrian as Points: An Improved Anchor-Free Method for Center-Based Pedestrian Detection [J].
Cai, Jiawei ;
Lee, Feifei ;
Yang, Shuai ;
Lin, Chaowei ;
Chen, Hanqing ;
Kotani, Koji ;
Chen, Qiu .
IEEE ACCESS, 2020, 8 :179666-179677
[4]   Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation [J].
Chen, Liang-Chieh ;
Zhu, Yukun ;
Papandreou, George ;
Schroff, Florian ;
Adam, Hartwig .
COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 :833-851
[5]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[6]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[7]  
Denton E, 2014, ADV NEUR IN, V27
[8]   DPP-Net: Device-Aware Progressive Search for Pareto-Optimal Neural Architectures [J].
Dong, Jin-Dong ;
Cheng, An-Chieh ;
Juan, Da-Cheng ;
Wei, Wei ;
Sun, Min .
COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 :540-555
[9]   Searching for A Robust Neural Architecture in Four GPU Hours [J].
Dong, Xuanyi ;
Yang, Yi .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :1761-1770
[10]  
Elsken T., 2019, PROC INT C LEARN REP