Sparse composition of body poses and atomic actions for human activity recognition in RGB-D videos

被引:42
作者
Lillo, Ivan [1 ]
Niebles, Juan Carlos [2 ,3 ]
Soto, Alvaro [1 ]
机构
[1] Pontificia Univ Catolica Chile, 4860 Vicuna Mackenna, Santiago, Chile
[2] Univ Norte, Barranquilla, Colombia
[3] Stanford Univ, Stanford, CA USA
关键词
Activity recognition; Hierarchical recognition model; RGB-D videos;
D O I
10.1016/j.imavis.2016.11.004
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents an approach to recognize human activities using body poses estimated from RGB-D data. We focus on recognizing complex activities composed of sequential or simultaneous atomic actions characterized by body motions of a single actor. We tackle this problem by introducing a hierarchical compositional model that operates at three levels of abstraction. At the lowest level, geometric and motion descriptors are used to learn a dictionary of body poses. At the intermediate level, sparse compositions of these body poses are used to obtain meaningful representations for atomic human actions. Finally, at the highest level, spatial and temporal compositions of these atomic actions are used to represent complex human activities. Our results show the benefits of using a hierarchical model that exploits the sharing and composition of body poses into atomic actions, and atomic actions into activities. A quantitative evaluation using two benchmark datasets illustrates the advantages of our model to perform action and activity recognition. (C) 2016 Elsevier B.V. All rights reserved.
引用
收藏
页码:63 / 75
页数:13
相关论文
共 42 条
[1]   Human activity recognition from 3D data: A review [J].
Aggarwal, J. K. ;
Xia, Lu .
PATTERN RECOGNITION LETTERS, 2014, 48 :70-80
[2]   Human Activity Analysis: A Review [J].
Aggarwal, J. K. ;
Ryoo, M. S. .
ACM COMPUTING SURVEYS, 2011, 43 (03)
[3]  
[Anonymous], 2011, IEEE Transactions on Visualization and Computer Graphics, DOI DOI 10.1109/TVCG.2010.272
[4]  
[Anonymous], KIN WIND SDK
[5]  
[Anonymous], IEEE C COMP VIS PATT
[6]  
[Anonymous], 2010, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2010.5539963
[7]  
Csurka G., 2004, WORKSH STAT LEARN CO, V1, P1, DOI DOI 10.1234/12345678
[8]  
Du Y, 2015, PROC CVPR IEEE, P1110, DOI 10.1109/CVPR.2015.7298714
[9]  
Escorcia Victor, 2012, Proceedings of the 2012 Construction Research Congress, P879
[10]  
Felzenszwalb P, 2008, PROC CVPR IEEE, P1984