Automatic Generation of Dynamic Inference Architecture for Deep Neural Networks

被引:1
作者
Zhao, Shize [1 ]
He, Liulu [1 ]
Xie, Xiaoru [1 ]
Lin, Jun [1 ]
Wang, Zhongfeng [1 ]
机构
[1] Nanjing Univ, Sch Elect Sci & Engn, Nanjing, Peoples R China
来源
2021 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2021) | 2021年
基金
中国国家自然科学基金;
关键词
Convolutional neural network (CNN); Neural architecture search (NAS); BranchyNet; Neighborhood greedy search;
D O I
10.1109/SiPS52927.2021.00029
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The computational cost of deep neural network(DNN) model can be reduced dramatically by applying different architectures based on the difficulties of each sample, which is named dynamic inference tech. Manually designed dynamic inference framework is hard to be optimal for the dependency on human experience, which is also time-consuming and labor-intensive. In this paper, we provide an auto-designed AB-Net based on the popular dynamic framework BranchyNet, which is inspired by neural architecture search (NAS). To further accelerate the search procedure, we also develop several specific techniques. Firstly, the search space is optimized by the pre-selection of candidate architectures. Then, a neighborhood greedy search algorithm is developed to efficiently find the optimal architecture in the improved search space. Moreover, our scheme can be extended to the multiple-branch cases to further enhance the performance of the AB-Net. We apply the AB-Net on multiple mainstream models and evaluate them on datasets CIFAR10/100. Compared to the handcrafted BranchyNet, the proposed AB-Net is able to achieve 1.57 x computational cost reduction at least even with slight accuracy improvement on CIFAR100. Moreover, the AB-Net also significantly outperforms the S2DNAS on accuracy with similar cost reduction, which is the state-of-the-art automatic dynamic inference architecture.
引用
收藏
页码:117 / 122
页数:6
相关论文
共 19 条
[1]  
Bender G, 2018, PR MACH LEARN RES, V80
[2]  
Bolukbasi T., 2017, PR MACH LEARN RES, P527
[3]   Spatially Adaptive Computation Time for Residual Networks [J].
Figurnov, Michael ;
Collins, Maxwell D. ;
Zhu, Yukun ;
Zhang, Li ;
Huang, Jonathan ;
Vetrov, Dmitry ;
Salakhutdinov, Ruslan .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1790-1799
[4]  
Graves, 2016, ADAPTIVE COMPUTATION
[5]  
Han S, 2015, ADV NEUR IN, V28
[6]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[7]  
Huang Gao, 2017, ARXIV PREPRINT ARXIV, P2
[8]  
Kaya Y, 2019, PR MACH LEARN RES, V97
[9]   Auto-DeepLab: Hierarchical Neural Architecture Search for Semantic Image Segmentation [J].
Liu, Chenxi ;
Chen, Liang-Chieh ;
Schroff, Florian ;
Adam, Hartwig ;
Hua, Wei ;
Yuille, Alan ;
Li Fei-Fei .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :82-92
[10]  
Liu H., 2017, ARXIV171100436