Feature Representations for Neuromorphic Audio Spike Streams

被引:72
作者
Anumula, Jithendar [1 ]
Neil, Daniel [2 ]
Delbruck, Tobi
Liu, Shih-Chii
机构
[1] Univ Zurich, Inst Neuroinformat, Zurich, Switzerland
[2] BenevolentAI, New York, NY USA
关键词
dynamic audio sensor; spike feature generation; exponential kernels; recurrent neural network; audio word classification; EVENT; NETWORKS; SENSOR; TIME;
D O I
10.3389/fnins.2018.00023
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
Event-driven neuromorphic spiking sensors such as the silicon retina and the silicon cochlea encode the external sensory stimuli as asynchronous streams of spikes across different channels or pixels. Combining state-of-art deep neural networks with the asynchronous outputs of these sensors has produced encouraging results on some datasets but remains challenging. While the lack of effective spiking networks to process the spike streams is one reason, the other reason is that the pre-processing methods required to convert the spike streams to frame-based features needed for the deep networks still require further investigation. This work investigates the effectiveness of synchronous and asynchronous frame-based features generated using spike count and constant event binning in combination with the use of a recurrent neural network for solving a classification task using N-TIDIGITS18 dataset. This spike-based dataset consists of recordings from the Dynamic Audio Sensor, a spiking silicon cochlea sensor, in response to the TIDIGITS audio dataset. We also propose a new pre-processing method which applies an exponential kernel on the output cochlea spikes so that the interspike timing information is better preserved. The results from the N-TIDIGITS18 dataset show that the exponential features perform better than the spike count features, with over 91% accuracy on the digit classification task. This accuracy corresponds to an improvement of at least 2.5% over the use of spike count features, establishing a new state of the art for this dataset.
引用
收藏
页数:12
相关论文
共 49 条
  • [1] ABADI M., 2015, TensorFlow: large-scale machine learning on heterogeneous systems
  • [2] Abdollahi M, 2011, BIOMED CIRC SYST C, P269, DOI 10.1109/BioCAS.2011.6107779
  • [3] Amir A., 2017, PROC CVPR IEEE, P7243, DOI DOI 10.1109/CVPR.2017.781
  • [4] [Anonymous], 2017, ADV NEURAL INFORM PR
  • [5] Mapping from Frame-Driven to Frame-Free Event-Driven Vision Systems by Low-Rate Rate Coding and Coincidence Processing-Application to Feedforward ConvNets
    Antonio Perez-Carrasco, Jose
    Zhao, Bo
    Serrano, Carmen
    Acha, Begona
    Serrano-Gotarredona, Teresa
    Chen, Shouchun
    Linares-Barranco, Bernabe
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (11) : 2706 - 2719
  • [6] Anumula J., 2017, P IEEE INT S CIRC SY, DOI [10.1109/ISCAS.2017.8050394, DOI 10.1109/ISCAS.2017.8050394]
  • [7] A Dataset for Visual Navigation with Neuromorphic Methods
    Barranco, Francisco
    Fermuller, Cornelia
    Aloimonos, Yiannis
    Delbruck, Tobi
    [J]. FRONTIERS IN NEUROSCIENCE, 2016, 10
  • [8] Berner Raphael, 2013, 2013 Symposium on VLSI Circuits, pC186
  • [9] Adaptive exponential integrate-and-fire model as an effective description of neuronal activity
    Brette, R
    Gerstner, W
    [J]. JOURNAL OF NEUROPHYSIOLOGY, 2005, 94 (05) : 3637 - 3642
  • [10] Chakrabartty S, 2010, IEEE INT SYMP CIRC S, P513, DOI 10.1109/ISCAS.2010.5537578