Universal Binary Neural Networks Design by Improved Differentiable Neural Architecture Search

被引:6
作者
Tan, Menghao [1 ]
Gao, Weifeng [1 ]
Li, Hong [1 ]
Xie, Jin [1 ]
Gong, Maoguo [2 ]
机构
[1] Xidian Univ, Sch Math & Stat, Xian 710126, Peoples R China
[2] Xidian Univ, Int Res Ctr Intelligent Percept & Computat, Key Lab Intelligent Percept & Image Understanding, Minist Educ, Xian 710071, Peoples R China
关键词
Computer architecture; Quantization (signal); Neural networks; Training; Microprocessors; Convolution; Circuits and systems; Binary neural networks; neural architecture search; search optimization;
D O I
10.1109/TCSVT.2024.3398691
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Binary Neural Networks (BNNs) using 1-bit weights and activations are emerging as a promising approach for mobile devices and edge computing platforms. Concurrently, traditional Neural Architecture Search (NAS) has gained widespread usage in automatically designing network architectures. However, the computation involved in binary NAS is more complex than in NAS due to the substantial information loss incurred by binary modules, and different binary spaces are required for different tasks. To address these challenges, a universal binary neural architecture search (UBNAS) algorithm is proposed. In this paper, the ApproxSign function is used to reduce the gradient error and accelerate the convergence in binary network searching and training. Moreover, UBNAS adopts a novel search space consisting of operations appropriate for the binary methods. To improve the original space operation module, we explore the effect of diverse structures for various modules and ultimately obtain a universal binary network structure. Additionally, the channel sampling ratio is adjusted to balance the advantages of different operations and an early stopping strategy is implemented to significantly reduce the computational burden associated with searching. We perform extensive experiments on CIFAR10, and ImageNet datasets and the results demonstrate the effectiveness of the proposed method.
引用
收藏
页码:9153 / 9165
页数:13
相关论文
共 70 条
[61]   Joint-Guided Distillation Binary Neural Network via Dynamic Channel-Wise Diversity Enhancement for Object Detection [J].
Xie, Yefan ;
Hou, Xuan ;
Guo, Yanwei ;
Wang, Xiuying ;
Zheng, Jiangbin .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (01) :448-460
[62]  
Xu YH, 2020, Arxiv, DOI arXiv:1907.05737
[63]   CDbin: Compact Discriminative Binary Descriptor Learned With Efficient Neural Network [J].
Ye, Jianming ;
Zhang, Shiliang ;
Huang, Tiejun ;
Rui, Yong .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (03) :862-874
[64]  
Zagoruyko S, 2017, Arxiv, DOI arXiv:1605.07146
[65]   Self-Growing Binary Activation Network: A Novel Deep Learning Model With Dynamic Architecture [J].
Zhang, Zeyang ;
Chen, Yidong ;
Zhou, Changle .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (01) :624-633
[66]  
Zhou BL, 2018, Arxiv, DOI arXiv:1806.02891
[67]  
Zhou S., 2016, arXiv
[68]   Towards Effective Low-bitwidth Convolutional Neural Networks [J].
Zhuang, Bohan ;
Shen, Chunhua ;
Tan, Mingkui ;
Liu, Lingqiao ;
Reid, Ian .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7920-7928
[69]  
Zhuo LA, 2020, PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P1033
[70]  
Zoph B, 2017, Arxiv, DOI arXiv:1611.01578