Spike representation of depth image sequences and its application to hand gesture recognition with spiking neural network

被引:3
作者
Miki, Daisuke [1 ]
Kamitsuma, Kento [1 ]
Matsunaga, Taiga [1 ]
机构
[1] Chiba Inst Technol, Dept Comp Sci, 2-17-1 Tsudanuma, Narashino, Chiba 2750016, Japan
基金
日本学术振兴会;
关键词
Spiking neural network; Hand gesture recognition; Depth image; Surrogate gradient learning;
D O I
10.1007/s11760-023-02574-3
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Hand gestures play an important role in expressing the emotions of people and communicating their intentions. Therefore, various methods have been studied to clearly capture and understand them. Artificial neural networks (ANNs) are widely used for gesture recognition owing to their expressive power and ease of implementation. However, this task remains challenging because it requires abundant data and energy for computation. Recently, low-power neuromorphic devices that use spiking neural networks (SNNs), which can process temporal information and require lower power consumption for computing, have attracted significant research interest. In this study, we present a method for the spike representation of human hand gestures and analyzing them using SNNs. An SNN comprises multiple convolutional layers; when a sequence of spike trains corresponding to a hand gesture is inputted, the spiking neurons in the output layer corresponding to each gesture fire, and the gesture is classified based on its firing frequency. Using a sequence of depth images of hand gestures, a method to generate spike trains from the training image data was investigated. The gestures could be classified by training the SNN using surrogate gradient (SG) learning. Additionally, by converting the depth image data into spike trains, 68% of the training data volume could be reduced without significantly reducing the classification accuracy, compared to the classification accuracy under ANNs.
引用
收藏
页码:3505 / 3513
页数:9
相关论文
共 48 条
[1]   Optuna: A Next-generation Hyperparameter Optimization Framework [J].
Akiba, Takuya ;
Sano, Shotaro ;
Yanase, Toshihiko ;
Ohta, Takeru ;
Koyama, Masanori .
KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, :2623-2631
[2]  
Amir Arnon, 2017, P IEEE C COMPUTER VI, P7388, DOI [DOI 10.1109/CVPR.2017.781, 10.1109/CVPR.2017.781]
[3]  
[Anonymous], 2022, GARTNER HYPE CYCLE
[4]  
[Anonymous], P IEEECVF C COMPUTER, DOI DOI 10.1109/CVPR.2016.456
[5]  
Bergstra J., 2011, Advances in neural information processing systems, P24
[6]   ExtriDeNet: an intensive feature extrication deep network for hand gesture recognition [J].
Bhaumik, Gopa ;
Verma, Monu ;
Govil, Mahesh Chandra ;
Vipparthi, Santosh Kumar .
VISUAL COMPUTER, 2022, 38 (11) :3853-3866
[7]  
Bi Y, 2017, IEEE IMAGE PROC, P1990, DOI 10.1109/ICIP.2017.8296630
[8]   Advancing Neuromorphic Computing With Loihi: A Survey of Results and Outlook [J].
Davies, Mike ;
Wild, Andreas ;
Orchard, Garrick ;
Sandamirskaya, Yulia ;
Guerra, Gabriel A. Fonseca ;
Joshi, Prasad ;
Plank, Philipp ;
Risbud, Sumedh R. .
PROCEEDINGS OF THE IEEE, 2021, 109 (05) :911-934
[9]   Skeleton-based Dynamic hand gesture recognition [J].
De Smedt, Quentin ;
Wannous, Hazem ;
Vandeborre, Jean-Philippe .
PROCEEDINGS OF 29TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, (CVPRW 2016), 2016, :1206-1214
[10]   Nengo and Low-Power AI Hardware for Robust, Embedded Neurorobotics [J].
DeWolf, Travis ;
Jaworski, Pawel ;
Eliasmith, Chris .
FRONTIERS IN NEUROROBOTICS, 2020, 14