SA-FlowNet: Event-based self-attention optical flow estimation with spiking-analogue neural networks

被引:6
作者
Yang, Fan [1 ]
Su, Li [1 ,3 ]
Zhao, Jinxiu [1 ]
Chen, Xuena [1 ]
Wang, Xiangyu [1 ]
Jiang, Na [1 ]
Hu, Quan [2 ]
机构
[1] Capital Normal Univ, Informat Engn Coll, Beijing, Peoples R China
[2] Beijing Inst Technol, Sch Aerosp Engn, Beijing, Peoples R China
[3] Capital Normal Univ, Informat Engn Coll, Beijing 100048, Peoples R China
基金
中国国家自然科学基金;
关键词
computer vision; feature extraction; motion estimation; optical tracking; INTELLIGENCE;
D O I
10.1049/cvi2.12206
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Inspired by biological vision mechanism, event-based cameras have been developed to capture continuous object motion and detect brightness changes independently and asynchronously, which overcome the limitations of traditional frame-based cameras. Complementarily, spiking neural networks (SNNs) offer asynchronous computations and exploit the inherent sparseness of spatio-temporal events. Notably, event-based pixel-wise optical flow estimations calculate the positions and relationships of objects in adjacent frames; however, as event camera outputs are sparse and uneven, dense scene information is difficult to generate and the local receptive fields of the neural network also lead to poor moving objects tracking. To address these issues, an improved event-based self-attention optical flow estimation network (SA-FlowNet) that independently uses criss-cross and temporal self-attention mechanisms, directly capturing long-range dependencies and efficiently extracting the temporal and spatial features from the event streams is proposed. In the former mechanism, a cross-domain attention scheme dynamically fusing the temporal-spatial features is introduced. The proposed network adopts a spiking-analogue neural network architecture using an end-to-end learning method and gains significant computational energy benefits especially for SNNs. The state-of-the-art results of the error rate for optical flow prediction on the Multi-Vehicle Stereo Event Camera (MVSEC) dataset compared with the current SNN-based approaches is demonstrated.
引用
收藏
页码:925 / 935
页数:11
相关论文
共 41 条
[1]  
[Anonymous], 2012, J CRIT CARE, DOI DOI 10.1016/J.JCRC.2012.01.039
[2]   PERFORMANCE OF OPTICAL-FLOW TECHNIQUES [J].
BARRON, JL ;
FLEET, DJ ;
BEAUCHEMIN, SS .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 1994, 12 (01) :43-77
[3]   Event-Based Visual Flow [J].
Benosman, Ryad ;
Clercq, Charles ;
Lagorce, Xavier ;
Ieng, Sio-Hoi ;
Bartolozzi, Chiara .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (02) :407-417
[4]   A 240 x 180 130 dB 3 μs Latency Global Shutter Spatiotemporal Vision Sensor [J].
Brandli, Christian ;
Berner, Raphael ;
Yang, Minhao ;
Liu, Shih-Chii ;
Delbruck, Tobi .
IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2014, 49 (10) :2333-2341
[5]   Lucas/Kanade meets Horn/Schunck: Combining local and global optic flow methods [J].
Bruhn A. ;
Weickert J. ;
Schnörr C. .
International Journal of Computer Vision, 2005, 61 (3) :1-21
[6]   A review of the integrate-and-fire neuron model: I. Homogeneous synaptic input [J].
Burkitt, A. N. .
BIOLOGICAL CYBERNETICS, 2006, 95 (01) :1-19
[7]   GCNet: Non-local Networks Meet Squeeze-Excitation Networks and Beyond [J].
Cao, Yue ;
Xu, Jiarui ;
Lin, Stephen ;
Wei, Fangyun ;
Hu, Han .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, :1971-1980
[8]  
Ding ZL, 2022, AAAI CONF ARTIF INTE, P525
[9]   Dual Attention Network for Scene Segmentation [J].
Fu, Jun ;
Liu, Jing ;
Tian, Haijie ;
Li, Yong ;
Bao, Yongjun ;
Fang, Zhiwei ;
Lu, Hanqing .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3141-3149
[10]   Event-Based Vision: A Survey [J].
Gallego, Guillermo ;
Delbruck, Tobi ;
Orchard, Garrick Michael ;
Bartolozzi, Chiara ;
Taba, Brian ;
Censi, Andrea ;
Leutenegger, Stefan ;
Davison, Andrew ;
Conradt, Jorg ;
Daniilidis, Kostas ;
Scaramuzza, Davide .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (01) :154-180