Automotive Object Detection via Learning Sparse Events by Spiking Neurons

被引:0
作者
Zhang, Hu [1 ]
Li, Yanchen [1 ]
Leng, Luziwei [2 ]
Che, Kaiwei [1 ]
Liu, Qian [2 ]
Guo, Qinghai [2 ]
Liao, Jianxing [2 ]
Cheng, Ran [1 ]
机构
[1] Southern Univ Sci & Technol, Shenzhen 518055, Peoples R China
[2] Huawei Technol Co Ltd, Adv Comp & Storage Lab, Shenzhen 518055, Peoples R China
关键词
Object detection; Task analysis; Neurons; Training; Vehicle dynamics; Feature extraction; Adaptation models; Deep learning; dynamical vision sensor (DVS); object detection; spiking neural networks (SNNs); NEURAL-NETWORKS; VISION;
D O I
10.1109/TCDS.2024.3410371
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Event-based sensors, distinguished by their high temporal resolution of 1 mu s and a dynamic range of 120 dB, stand out as ideal tools for deployment in fast-paced settings such as vehicles and drones. Traditional object detection techniques that utilize artificial neural networks (ANNs) face challenges due to the sparse and asynchronous nature of the events these sensors capture. In contrast, spiking neural networks (SNNs) offer a promising alternative, providing a temporal representation that is inherently aligned with event-based data. This article explores the unique membrane potential dynamics of SNNs and their ability to modulate sparse events. We introduce an innovative spike-triggered adaptive threshold mechanism designed for stable training. Building on these insights, we present a specialized spiking feature pyramid network (SpikeFPN) optimized for automotive event-based object detection. Comprehensive evaluations demonstrate that SpikeFPN surpasses both traditional SNNs and advanced ANNs enhanced with attention mechanisms. Evidently, SpikeFPN achieves a mean average precision (mAP) of 0.477 on the GEN1 automotive detection (GAD) benchmark dataset, marking significant increases over the selected SNN baselines. Moreover, the efficient design of SpikeFPN ensures robust performance while optimizing computational resources, attributed to its innate sparse computation capabilities.
引用
收藏
页码:2110 / 2124
页数:15
相关论文
共 99 条
  • [1] Real-Time 6DOF Pose Relocalization for Event Cameras with Stacked Spatial LSTM Networks
    Anh Nguyen
    Thanh-Toan Do
    Caldwell, Darwin G.
    Tsagarakis, Nikos G.
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, : 1638 - 1645
  • [2] Bellec G, 2018, ADV NEUR IN, V31
  • [3] A solution to the learning dilemma for recurrent networks of spiking neurons
    Bellec, Guillaume
    Scherr, Franz
    Subramoney, Anand
    Hajek, Elias
    Salaj, Darjan
    Legenstein, Robert
    Maass, Wolfgang
    [J]. NATURE COMMUNICATIONS, 2020, 11 (01)
  • [4] A 240 x 180 130 dB 3 μs Latency Global Shutter Spatiotemporal Vision Sensor
    Brandli, Christian
    Berner, Raphael
    Yang, Minhao
    Liu, Shih-Chii
    Delbruck, Tobi
    [J]. IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2014, 49 (10) : 2333 - 2341
  • [5] M3D-RPN: Monocular 3D Region Proposal Network for Object Detection
    Brazil, Garrick
    Liu, Xiaoming
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 9286 - 9295
  • [6] Bu T., 2022, P 10 INT C LEARN REP
  • [7] Cascade R-CNN: Delving into High Quality Object Detection
    Cai, Zhaowei
    Vasconcelos, Nuno
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 6154 - 6162
  • [8] Cannici Marco, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12365), P136, DOI 10.1007/978-3-030-58565-5_9
  • [9] Che KW, 2022, ADV NEUR IN
  • [10] A Novel Video Salient Object Detection Method via Semisupervised Motion Quality Perception
    Chen, Chenglizhao
    Song, Jia
    Peng, Chong
    Wang, Guodong
    Fang, Yuming
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (05) : 2732 - 2745