Simultaneous Learning Intensity and Optical Flow From High-Speed Spike Stream

被引:0
作者
Zhu, Lin [1 ]
Yan, Weiquan [2 ]
Chang, Yi [3 ]
Tian, Yonghong [4 ]
Huang, Hua [5 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing 100081, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[3] Huazhong Univ Sci & Technol, Sch Artificial Intelligence & Automat, Wuhan 430074, Peoples R China
[4] Peking Univ, Sch Comp Sci, Beijing 100081, Peoples R China
[5] Beijing Normal Univ, Sch Artificial Intelligence, Beijing 100875, Peoples R China
基金
中国国家自然科学基金;
关键词
Image reconstruction; Optical flow; Streaming media; Cameras; Feature extraction; Estimation; Vision sensors; Neurons; Image motion analysis; Decoding; optical flow estimation; bio-inspired camera; spiking neural network; EVENT CAMERAS; RECONSTRUCTION; NETWORKS;
D O I
10.1109/TCSVT.2024.3516478
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Bio-inspired vision sensors, which emulate the human retina by recording light intensity as binary spikes, have gained increasing interest in recent years. Among them, the spike camera is capable of perceiving fine textures by simulating a small retinal region called the fovea and producing high temporal resolution (20,000 Hz) spatiotemporal spike streams. To bridge the gap between binary spike streams and human vision in high-speed scenes, reconstructing intensity and optical flow from high temporal resolution spikes is particularly important. In this paper, we present a hybrid SNN-ANN network designed for simultaneous intensity and optical flow learning from spike streams. To adaptively extract spatial and temporal features from continuous spike streams, we propose a spiking neuron module with dense connections that efficiently processes both short-term and long-term spike data, while maintaining low power consumption characteristics. Subsequently, we introduce two decoders for optical flow and intensity estimation that complement each other. A temporal-aware warping module, based on flow features, is specifically designed to align the temporal features of the intensity decoder, thereby reducing motion artifacts. Concurrently, improved intensity features contribute to more accurate flow feature predictions, resulting in a mutually beneficial relationship within our network. To evaluate the effectiveness of our proposed network, we conduct experiments on both simulated and real spike datasets. Our network outperforms existing state-of-the-art spike-based reconstruction and optical flow estimation methods, demonstrating its potential for advancing the field of bio-inspired vision sensors. Our code is available at https://github.com/LinZhu111/SLIO.
引用
收藏
页码:5126 / 5139
页数:14
相关论文
共 57 条
[1]  
Chen SY, 2023, AAAI CONF ARTIF INTE, P350
[2]  
Chen SY, 2022, PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, P2859
[3]   Arbitrated time-to-first spike CMOS image sensor with on-chip histogram equalization [J].
Chen Shoushun ;
Bermak, Amine .
IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2007, 15 (03) :346-357
[4]  
Chen Sitan, 2023, ADV NEURAL INFORM PR
[5]   A biomorphic digital image sensor [J].
Culurciello, E ;
Etienne-Cummings, R ;
Boahen, KA .
IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2003, 38 (02) :281-294
[6]  
Delbrück T, 2010, IEEE INT SYMP CIRC S, P2426, DOI 10.1109/ISCAS.2010.5537149
[7]   Spike Camera and Its Coding Methods [J].
Dong, Siwei ;
Huang, Tiejun ;
Tian, Yonghong .
2017 DATA COMPRESSION CONFERENCE (DCC), 2017, :437-437
[8]   SpikingJelly: An open-source machine learning infrastructure platform for spike-based intelligence [J].
Fang, Wei ;
Chen, Yanqi ;
Ding, Jianhao ;
Yu, Zhaofei ;
Masquelier, Timothee ;
Chen, Ding ;
Huang, Liwei ;
Zhou, Huihui ;
Li, Guoqi ;
Tian, Yonghong .
SCIENCE ADVANCES, 2023, 9 (40)
[9]   Incorporating Learnable Membrane Time Constant to Enhance Learning of Spiking Neural Networks [J].
Fang, Wei ;
Yu, Zhaofei ;
Chen, Yanqi ;
Masquelier, Timothee ;
Huang, Tiejun ;
Tian, Yonghong .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :2641-2651
[10]   E-RAFT: Dense Optical Flow from Event Cameras [J].
Gehrig, Mathias ;
Millhaeusler, Mario ;
Gehrig, Daniel ;
Scaramuzza, Davide .
2021 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2021), 2021, :197-206