Energy-Efficient Spiking Segmenter for Frame and Event-Based Images

被引:10
作者
Zhang, Hong [1 ]
Fan, Xiongfei [1 ]
Zhang, Yu [1 ,2 ]
机构
[1] Zhejiang Univ, Coll Control Sci & Engn, State Key Lab Ind Control Technol, Hangzhou 310027, Peoples R China
[2] Key Lab Collaborat Sensing & Autonomous Unmanned S, Hangzhou 310027, Peoples R China
关键词
neuromophic computing; spiking neural network; semantic segmentation; spiking context guided network; frame and event-based images; NEURAL-NETWORKS;
D O I
10.3390/biomimetics8040356
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Semantic segmentation predicts dense pixel-wise semantic labels, which is crucial for autonomous environment perception systems. For applications on mobile devices, current research focuses on energy-efficient segmenters for both frame and event-based cameras. However, there is currently no artificial neural network (ANN) that can perform efficient segmentation on both types of images. This paper introduces spiking neural network (SNN, a bionic model that is energy-efficient when implemented on neuromorphic hardware) and develops a Spiking Context Guided Network (Spiking CGNet) with substantially lower energy consumption and comparable performance for both frame and event-based images. First, this paper proposes a spiking context guided block that can extract local features and context information with spike computations. On this basis, the directly-trained SCGNet-S and SCGNet-L are established for both frame and event-based images. Our method is verified on the frame-based dataset Cityscapes and the event-based dataset DDD17. On the Cityscapes dataset, SCGNet-S achieves comparable results to ANN CGNet with 4.85 x energy efficiency. On the DDD17 dataset, Spiking CGNet outperforms other spiking segmenters by a large margin.
引用
收藏
页数:18
相关论文
共 67 条
[41]   HyperSeg: Patch-wise Hypernetwork for Real-time Semantic Segmentation [J].
Nirkin, Yuval ;
Wolf, Lior ;
Hassner, Tal .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :4060-4069
[42]  
Paszke A, 2016, Arxiv, DOI arXiv:1606.02147
[43]   Full-Resolution Residual Networks for Semantic Segmentation in Street Scenes [J].
Pohlen, Tobias ;
Hermans, Alexander ;
Mathias, Markus ;
Leibe, Bastian .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :3309-3318
[44]  
Romera E, 2017, IEEE INT VEH SYM, P1789, DOI 10.1109/IVS.2017.7995966
[45]   U-Net: Convolutional Networks for Biomedical Image Segmentation [J].
Ronneberger, Olaf ;
Fischer, Philipp ;
Brox, Thomas .
MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION, PT III, 2015, 9351 :234-241
[46]   Conversion of Continuous-Valued Deep Networks to Efficient Event-Driven Networks for Image Classification [J].
Rueckauer, Bodo ;
Lungu, Iulia-Alexandra ;
Hu, Yuhuang ;
Pfeiffer, Michael ;
Liu, Shih-Chii .
FRONTIERS IN NEUROSCIENCE, 2017, 11
[47]   Going Deeper in Spiking Neural Networks: VGG and Residual Architectures [J].
Sengupta, Abhronil ;
Ye, Yuting ;
Wang, Robert ;
Liu, Chiao ;
Roy, Kaushik .
FRONTIERS IN NEUROSCIENCE, 2019, 13
[48]  
Shrestha SB, 2018, ADV NEUR IN, V31
[49]  
Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
[50]   Segmenter: Transformer for Semantic Segmentation [J].
Strudel, Robin ;
Garcia, Ricardo ;
Laptev, Ivan ;
Schmid, Cordelia .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :7242-7252