LCH: fast RGB-D salient object detection on CPU via lightweight convolutional network with hybrid knowledge distillation

被引:4
作者
Wang, Binglu [1 ,2 ]
Zhang, Fan [1 ]
Zhao, Yongqiang [1 ]
机构
[1] Northwestern Polytech Univ, Sch Automat, 127 West Youyi Rd, Xian, Shaanxi, Peoples R China
[2] Beijing Inst Technol, Sch Informat & Elect, 5 South St, Beijing 100081, Peoples R China
关键词
RGB-D SOD; Lightweight neural network; Knowledge distillation; Non-local network; FUSION;
D O I
10.1007/s00371-023-02898-8
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Recently, the performance of RGB-D salient object detection (SOD) has been significantly improved with the development of deep-learning techniques. However, most of them depend on complex structures with a large amount of parameters and multi-add operations, which need huge computational resources and are not applicable to real-world applications, especially on mobile devices. To handle this problem, we propose a lightweight network, namely LCH, for RGB-D SOD. Specifically, we first design a novel lightweight ghost in ghost (GiG) module to enable a more powerful multi-scale feature extraction, which not only efficiently tackles the scale variations of salient objects but also guarantees a high inference speed. Then, the proposed GiG module works in conjunction with ghost bottlenecks (G-bneck) and, therefore, forms a new lightweight U-Net structure network for fast CPU RGB-D SOD. Furthermore, to enhance the feature discriminability of our LCH, we also design a novel hybrid knowledge distillation approach (HKD), for which both semantic structure information and pixel-wise similarity information are transferred from a powerful state-of-the-art RGB-D SOD network to the proposed LCH model, so as to generate more satisfactory detection results. Compared with previous work, LCH owns 5.6x fewer parameters and achieves 5.4x faster inference speed on the CPU devices, but obtains comparable SOD performance. Extensive experiments are carried out on commonly-used RGB-D SOD datasets. Both qualitative and quantitative results prove the effectiveness of our proposed method.
引用
收藏
页码:1997 / 2014
页数:18
相关论文
共 96 条
[1]   Efficient Semantic Segmentation via Self-Attention and Self-Distillation [J].
An, Shumin ;
Liao, Qingmin ;
Lu, Zongqing ;
Xue, Jing-Hao .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (09) :15256-15266
[2]   Salient Object Detection: A Benchmark [J].
Borji, Ali ;
Sihite, Dicky N. ;
Itti, Laurent .
COMPUTER VISION - ECCV 2012, PT II, 2012, 7573 :414-429
[3]  
Chen GB, 2017, ADV NEUR IN, V30
[4]   Three-Stream Attention-Aware Network for RGB-D Salient Object Detection [J].
Chen, Hao ;
Li, Youfu .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (06) :2825-2835
[5]   Discriminative Cross-Modal Transfer Learning and Densely Cross-Level Feedback Fusion for RGB-D Salient Object Detection [J].
Chen, Hao ;
Li, Youfu ;
Su, Dan .
IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (11) :4808-4820
[6]   Progressively Complementarity-aware Fusion Network for RGB-D Salient Object Detection [J].
Chen, Hao ;
Li, Youfu .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :3051-3060
[7]   Multi-modal fusion network with multi-scale multi-path and cross-modal interactions for RGB-D salient object detection [J].
Chen, Hao ;
Li, Youfu ;
Su, Dan .
PATTERN RECOGNITION, 2019, 86 :376-385
[8]   DISC: Deep Image Saliency Computing via Progressive Representation Learning [J].
Chen, Tianshui ;
Lin, Liang ;
Liu, Lingbo ;
Luo, Xiaonan ;
Li, Xuelong .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (06) :1135-1149
[9]  
Chen Y., 2017, arXiv
[10]   Mobile-Former: Bridging MobileNet and Transformer [J].
Chen, Yinpeng ;
Dai, Xiyang ;
Chen, Dongdong ;
Liu, Mengchen ;
Dong, Xiaoyi ;
Yuan, Lu ;
Liu, Zicheng .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :5260-5269