FBNetV2: Differentiable Neural Architecture Search for Spatial and Channel Dimensions

被引:200
作者
Wan, Alvin [1 ,3 ]
Dai, Xiaoliang [2 ]
Zhang, Peizhao [2 ]
He, Zijian [2 ]
Tian, Yuandong [2 ]
Xie, Saining [2 ]
Wu, Bichen [2 ]
Yu, Matthew [2 ]
Xu, Tao [2 ]
Chen, Kan [2 ]
Vajda, Peter [2 ]
Gonzalez, Joseph E. [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
[2] Facebook Inc, Menlo Pk, CA USA
[3] Facebook, Menlo Pk, CA USA
来源
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020) | 2020年
关键词
D O I
10.1109/CVPR42600.2020.01298
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Differentiable Neural Architecture Search (DNAS) has demonstrated great success in designing state-of-the-art, efficient neural networks. However, DARTS-based DNAS's search space is small when compared to other search methods', since all candidate network layers must be explicitly instantiated in memory. To address this bottleneck, we propose a memory and computationally efficient DNAS variant: DMaskingNAS. This algorithm expands the search space by up to 10(14)x over conventional DNAS, supporting searches over spatial and channel dimensions that are otherwise prohibitively expensive: input resolution and number of filters. We propose a masking mechanism for feature map reuse, so that memory and computational costs stay nearly constant as the search space expands. Furthermore, we employ effective shape propagation to maximize per-FLOP or per-parameter accuracy. The searched FBNetV2s yield state-of-the-art performance when compared with all previous architectures. With up to 421x less search cost, DMaskingNAS finds models with 0.9% higher accuracy, 15% fewer FLOPs than MobileNetV3-Small; and with similar accuracy but 20% fewer FLOPs than Efficient-B0. Furthermore, our FBNetV2 outperforms MobileNetV3 by 2.6% in accuracy, with equivalent model size. FBNetV2 models are open-sourced at https://github.com/facebookresearch/mobile-vision.
引用
收藏
页码:12962 / 12971
页数:10
相关论文
共 40 条
[1]  
[Anonymous], 2018, Energy
[2]  
Cai Han, 2018, ARXIV
[3]   ChamNet: Towards Efficient Network Design through Platform-Aware Model Adaptation [J].
Dai, Xiaoliang ;
Zhang, Peizhao ;
Wu, Bichen ;
Yin, Hongxu ;
Sun, Fei ;
Wang, Yanghan ;
Dukhan, Marat ;
Hu, Yunqing ;
Wu, Yiming ;
Jia, Yangqing ;
Vajda, Peter ;
Uyttendaele, Matt ;
Jha, Niraj K. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11390-11399
[4]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[5]  
docs.nvidia, DEEP LEARN PERF GUID
[6]  
Elsken T, 2019, J MACH LEARN RES, V20
[7]  
Guo ZJ, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P241
[8]  
Han S., 2016, INT C LEARN REPR ICL
[9]  
Han S., 2015, ARXIV PREPRINT ARXIV
[10]  
HE KM, 2016, PROC CVPR IEEE, P770, DOI DOI 10.1109/CVPR.2016.90