FreeNet: An efficient frequency-domain early exiting network for dynamic inference

被引:0
作者
Yan, Yicheng [1 ]
Li, Xianfeng [1 ]
Cui, Kai [2 ]
Sun, Haoran [1 ]
Yu, Zifeng [1 ]
机构
[1] Macau Univ Sci & Technol, Fac Innovat Engn, Taipa 999078, Macao, Peoples R China
[2] Zhejiang Univ, Innovat Ctr Yangtze River Delta, Jiaxing 314121, Peoples R China
关键词
Dynamic neural network; Efficient inference; Early exiting; Frequency domain; Image recognition; NEURAL-NETWORK;
D O I
10.1016/j.knosys.2025.113155
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Early exiting has become an effective method for improving inference efficiency. Most early exiting models place early classifiers in shallow layers, forcing reliance on low-level features and limiting performance. Moreover, these models typically use the spatial representation of the image (RGB image) as input, reducing spatial redundancy only by lowering the resolution. Compared to spatial representation, frequency domain representation provides richer multilevel information, enabling more efficient early exiting mechanisms. To this end, this paper proposes a dynamic frequency-domain early exiting network (FreeNet). Our model employs wavelet transform to decompose images into multiple levels of frequency information, with distinct frequency sub-models dedicated to feature extraction at each level. Each classifier is positioned at the end of its frequency sub-model, ensuring it can leverage high-level features with semantic information. Our model performs inference sequentially from low frequency to high frequency. For easy samples, it uses only a small amount of low-frequency information, while for hard samples, it dynamically supplements high-frequency information to ensure reliable results. Extensive experiments applying our method to various baseline models show that our model achieves abetter balance between performance and computational cost. For instance, on the ImageNet-1k dataset, our method can reduce 58% FLOPs of ResNet50 by only a 0.3% accuracy loss.
引用
收藏
页数:11
相关论文
共 36 条
[1]   Dissecting the High-Frequency Bias in Convolutional Neural Networks [J].
Abello, Antonio A. ;
Hirata Jr, Roberto ;
Wang, Zhangyang .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, :863-871
[2]  
Bolukbasi T, 2017, PR MACH LEARN RES, V70
[3]   Drop an Octave: Reducing Spatial Redundancy in Convolutional Neural Networks with Octave Convolution [J].
Chen, Yunpeng ;
Fan, Haoqi ;
Xu, Bing ;
Yan, Zhicheng ;
Kalantidis, Yannis ;
Rohrbach, Marcus ;
Yan, Shuicheng ;
Feng, Jiashi .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3434-3443
[4]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[5]   DepGraph: Towards Any Structural Pruning [J].
Fang, Gongfan ;
Ma, Xinyin ;
Song, Mingli ;
Mi, Michael Bi ;
Wang, Xinchao .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, :16091-16101
[6]  
Howard AG, 2017, Arxiv, DOI [arXiv:1704.04861, 10.48550/arXiv.1704.04861]
[7]  
Gueguen L, 2018, ADV NEUR IN, V31
[8]   Sensitivity pruner: Filter-Level compression algorithm for deep neural networks [J].
Guo, Suhan ;
Lai, Bilan ;
Yang, Suorong ;
Zhao, Jian ;
Shen, Furao .
PATTERN RECOGNITION, 2023, 140
[9]   Learning to Weight Samples for Dynamic Early-Exiting Networks [J].
Han, Yizeng ;
Pu, Yifan ;
Lai, Zihang ;
Wang, Chaofei ;
Song, Shiji ;
Cao, Junfeng ;
Huang, Wenhui ;
Deng, Chao ;
Huang, Gao .
COMPUTER VISION, ECCV 2022, PT XI, 2022, 13671 :362-378
[10]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778