CSTR: A Compact Spatio-Temporal Representation for Event-Based Vision

被引:3
作者
El Shair, Zaid A. [1 ]
Hassani, Ali [1 ]
Rawashdeh, Samir A. [1 ]
机构
[1] Univ Michigan, Dept Elect & Comp Engn, Dearborn, MI 48128 USA
关键词
Chemical reactions; Computer vision; Computer architecture; Task analysis; Vehicle dynamics; Object recognition; Visualization; Event detection; Object detection; Data augmentation; Event-based vision; event representation; object recognition; data augmentation; LATENCY;
D O I
10.1109/ACCESS.2023.3316143
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Event-based vision is a novel perception modality that offers several advantages, such as high dynamic range and robustness to motion blur. In order to process events in batches and utilize modern computer vision deep-learning architectures, an intermediate representation is required. Nevertheless, constructing an effective batch representation is non-trivial. In this paper, we propose a novel representation for event-based vision, called the compact spatio-temporal representation (CSTR). The CSTR encodes an event batch's spatial, temporal, and polarity information in a 3-channel image-like format. It achieves this by calculating the mean of the events' timestamps in combination with the event count at each spatial position in the frame. This representation shows robustness to motion-overlapping, high event density, and varying event-batch durations. Due to its compact 3-channel form, the CSTR is directly compatible with modern computer vision architectures, serving as an excellent choice for deploying event-based solutions. In addition, we complement the CSTR with an augmentation framework that introduces randomized training variations to the spatial, temporal, and polarity characteristics of event data. Experimentation over different object and action recognition datasets shows that the CSTR outperforms other representations of similar complexity under a consistent baseline. Further, the CSTR is made more robust and significantly benefits from the proposed augmentation framework, considerably addressing the sparseness in event-based datasets.
引用
收藏
页码:102899 / 102916
页数:18
相关论文
共 63 条
  • [1] EV-SegNet: Semantic Segmentation for Event-based Cameras
    Alonso, Inigo
    Murillo, Ana C.
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, : 1624 - 1633
  • [2] Amir A., P IEEE C COMP VIS PA
  • [3] Accurate and Efficient Frame-based Event Representation for AER Object Recognition
    Bai, WeiJie
    Chen, Yunhua
    Feng, Ren
    Zheng, Yuliang
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [4] Time-Ordered Recent Event (TORE) Volumes for Event Cameras
    Baldwin, R. Wes
    Liu, Ruixu
    Almatrafi, Mohammed
    Asari, Vijayan
    Hirakawa, Keigo
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (02) : 2519 - 2532
  • [5] Benenson R., 2008, International Journal of Vehicle Autonomous Systems, Inderscience, V1, P4
  • [6] Graph-Based Spatio-Temporal Feature Learning for Neuromorphic Vision Sensing
    Bi, Yin
    Chadha, Aaron
    Abbas, Alhabib
    Bourtsoulatze, Eirina
    Andreopoulos, Yiannis
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 9084 - 9098
  • [7] Graph-Based Object Classification for Neuromorphic Vision Sensing
    Bi, Yin
    Chadha, Aaron
    Abbas, Alhabib
    Bourtsoulatze, Eirina
    Andreopoulos, Yiannis
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 491 - 501
  • [8] Human-Robot Perception in Industrial Environments: A Survey
    Bonci, Andrea
    Cen Cheng, Pangcheng David
    Indri, Marina
    Nabissi, Giacomo
    Sibona, Fiorella
    [J]. SENSORS, 2021, 21 (05) : 1 - 29
  • [9] A 240 x 180 130 dB 3 μs Latency Global Shutter Spatiotemporal Vision Sensor
    Brandli, Christian
    Berner, Raphael
    Yang, Minhao
    Liu, Shih-Chii
    Delbruck, Tobi
    [J]. IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2014, 49 (10) : 2333 - 2341
  • [10] Event-Based Neuromorphic Vision for Autonomous Driving: A Paradigm Shift for Bio-Inspired Visual Sensing and Perception
    Chen, Guang
    Cao, Hu
    Conradt, Jorg
    Tang, Huajin
    Rohrbein, Florian
    Knoll, Alois
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2020, 37 (04) : 34 - 49