Automatic Action Segmentation and Continuous Recognition for Basic Indoor Actions Based on Kinect Pose Streams

被引:0
作者
Han, Yun [1 ]
Chung, Sheng-Luen [2 ]
Su, Shun-Feng [2 ]
机构
[1] Neijiang Normal Univ, Sch Comp Sci, Neijiang, Sichuan, Peoples R China
[2] Natl Taiwan Univ Sci & Technol, Dept Elect Engn, Taipei, Taiwan
来源
2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC) | 2017年
关键词
action recognition; on-line monitoring; Kinect;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One main difficulty in applying action recognition to practical applications is the need to segment beginnings and ends of actions in a continuous online monitoring process. This paper proposed a finite state machine (FSM) model for automatic action segmentation and recognition solution, based on pose streams in the form of skeleton joint data provided by Kinect. With the action recognition problem reframed as a state identification problem, the key solution to state identification hinges on detection of changing events, which signify the start of new action and the recognition of the underlying action. In that regard, a decision tree is constructed to detect these events based on the spatial positions and the changes of the skeleton data. In addition, to identify the current state or equivalently the action of a detected person, a fault observer is derived from the modeled action FSM. The fault observer does not only identify the initial state of action when the recognition process starts, but also serve the purpose of error recovery when the system loses track of ongoing events at times of intermittent sensor faults. Additionally, an AutoCorrect mechanism is presented to further enhance the accuracy of action recognition. To evaluate the proposed approach, an experiment with 300 participating subjects has been conducted for a total of 900 test sequences. The 98.63% correct identification result ensures the proposed approach a promising solution to constant action monitoring solution.
引用
收藏
页码:966 / 971
页数:6
相关论文
共 18 条
[1]  
[Anonymous], P 2016 ACM MUL TIM C
[2]  
[Anonymous], 2013, CMU GRAPH LAB MOT CA
[3]  
[Anonymous], 2012, IEEE COMP SOC C COMP, DOI DOI 10.1109/CVPRW.2012.6239234
[4]  
Cassandras C. G., 2009, INTRO DISCRETE EVENT, DOI DOI 10.1007/978-3-030-72274-6
[5]   Real Time Motion Capture Using a Single Time-Of-Flight Camera [J].
Ganapathi, Varun ;
Plagemann, Christian ;
Koller, Daphne ;
Thrun, Sebastian .
2010 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2010, :755-762
[6]   Local polynomial space-time descriptors for action classification [J].
Kihl, Olivier ;
Picard, David ;
Gosselin, Philippe-Henri .
MACHINE VISION AND APPLICATIONS, 2016, 27 (03) :351-361
[7]   Enhanced skeleton visualization for view invariant human action recognition [J].
Liu, Mengyuan ;
Liu, Hong ;
Chen, Chen .
PATTERN RECOGNITION, 2017, 68 :346-362
[8]   Human Action Recognition of Hidden Markov Model Based on Depth Information [J].
Lu, Tongwei ;
Peng, Ling ;
Miao, Shaojun .
2016 15TH INTERNATIONAL SYMPOSIUM ON PARALLEL AND DISTRIBUTED COMPUTING (ISPDC), 2016, :354-357
[9]  
Ogale AbhijitS., 2004, In VACE
[10]   View-independent human action recognition with Volume Motion Template on single stereo camera [J].
Roh, Myung-Cheol ;
Shin, Ho-Keun ;
Lee, Seong-Whan .
PATTERN RECOGNITION LETTERS, 2010, 31 (07) :639-647