Feature distribution guided binary neural networks

被引:0
作者
Liu C. [1 ]
Chen Y. [1 ]
机构
[1] Key Laboratory of Advanced Process Control for Light Industry of Ministry of Education, Jiangnan University, Wuxi
来源
Kongzhi yu Juece/Control and Decision | 2024年 / 39卷 / 06期
关键词
binary neural networks; feature distribution; mean and variance adjustment; model compression; neural network quantization; semantic information speicherung;
D O I
10.13195/j.kzyjc.2022.1945
中图分类号
学科分类号
摘要
In recent years, binary neural networks (BNNs) have received attention due to their small memory consumption and high computational efficiency. However, there exists a significant performance gap between BNNs and floating-point deep neural networks (DNNs) due to problems, such as imbalanced distributions of positive and negative parts of quantized activation features, which affects their deployment on resource-constrained platforms. The main reason for the limited accuracy of binary networks is the information loss caused by feature discretization and the disappearance of semantic information caused by improper distribution optimization. To address this problem, this paper applies feature distribution adjustment to guide binarization, which adjusts the mean-variance of features to balance the feature distribution and reduce the information loss caused by discretization. At the same time, through the design of group excitation and feature fine-tuning module, the quantization zero points are optimized to balance the binarization activation distributions and retain the semantic information to the maximum extent. Experiments show that the proposed method achieves better results on different backbone networks using different datasets, in which only 0.4 % of accuracy is lost after binarizing ResNet-18 on CIFAR-10, which surpasses the current mainstream BNNs. © 2024 Northeast University. All rights reserved.
引用
收藏
页码:1840 / 1848
页数:8
相关论文
共 24 条
[1]  
Zheng X P, Liang X., Deep capsule network based on pruning optimization, Chinese Journal of Computers, 45, 7, pp. 1557-1570, (2022)
[2]  
Liu Z, Wang Y, Han K, Et al., Instance-aware dynamic neural network quantization, Proceedings of the Conference on Computer Vision and Pattern Recognition, pp. 12434-12443, (2022)
[3]  
Pan R D, Kong W J, Qi J., Legal judgment prediction based on pre-training model and knowledge distillation, Control and Decision, 37, 1, pp. 67-76, (2022)
[4]  
Cheng Q, Li J, Gao X L, Et al., Lightweight method of deep neural network based on deep sparse low rank decomposition, Control and Decision, 38, 3, pp. 751-758, (2023)
[5]  
Yu W Y, Zhang Y, Yao H M, Et al., Visual inspection of surface defects based on lightweight reconstruction network, Acta Automatica Sinica, 48, 9, pp. 2175-2186, (2022)
[6]  
Howard A G, Zhu M L, Chen B, Et al., MobileNets: Efficient convolutional neural networks for mobile vision applications, (2017)
[7]  
Courbariaux M, Bengio Y, David J P., Binary connect: Training deep neural networks with binary weights during propagations, (2015)
[8]  
Rastegari M, Ordonez V, Redmon J, Et al., XNOR-net: ImageNet classification using binary convolutional neural networks, Computer Vision — ECCV 2016, pp. 525-542, (2016)
[9]  
Qin H T, Gong R H, Liu X L, Et al., Forward and backward information retention for accurate binary neural networks, IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2250-2259, (2020)
[10]  
Xue P, Lu Y, Chang J F, Et al., Self-distribution binary neural networks, Applied Intelligence, 52, 12, pp. 13870-13882, (2022)