Action Status Based Novel Relative Feature Representations for Interaction Recognition

被引:10
作者
Li Yanshan [1 ,2 ]
Guo Tianyu [1 ,2 ]
Liu Xing [1 ,2 ]
Luo Wenhan [3 ]
Xie Weixin [1 ,2 ]
机构
[1] Shenzhen Univ, ATR Natl Key Lab Def Technol, Shenzhen 518000, Peoples R China
[2] Shenzhen Univ, Guangdong Key Lab Intelligent Informat Proc, Shenzhen 518000, Peoples R China
[3] Tencent, Shenzhen 518000, Peoples R China
基金
中国国家自然科学基金;
关键词
Human action analysis; Interaction recognition; Action status; Multi-stream network; Relative feature representations;
D O I
10.1049/cje.2020.00.088
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Skeleton-based action recognition has always been an important research topic in computer vision. Most of the researchers in this field currently pay more attention to actions performed by a single person while there is very little work dedicated to the identification of interactions between two people. However, the practical application of interaction recognition is actually more critical in our society considering that actions are often performed by multiple people. How to design an effective scheme to learn discriminative spatial and temporal representations for skeleton-based interaction recognition is still a challenging problem. Focusing on the characteristics of skeleton data for interactions, we first define the moving distance to distinguish the action status of the participants. Then some view-invariant relative features are proposed to fully represent the spatial and temporal relationship of the skeleton sequence. Further, a new coding method is proposed to obtain the novel relative feature representations. Finally, we design a three-stream CNN model to learn deep features for interaction recognition. We evaluate our method on SBU dataset, NTU RGB+D 60 dataset and NTU RGB+D 120 dataset. The experimental results also verify that our method is effective and exhibits great robustness compared with current state-of-the-art methods.
引用
收藏
页码:168 / 180
页数:13
相关论文
共 42 条
[1]   Human activity recognition from 3D data: A review [J].
Aggarwal, J. K. ;
Xia, Lu .
PATTERN RECOGNITION LETTERS, 2014, 48 :70-80
[2]   One of a Kind: Inferring Personality Impressions in Meetings [J].
Aran, Oya ;
Gatica-Perez, Daniel .
ICMI'13: PROCEEDINGS OF THE 2013 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2013, :11-18
[3]  
Aydin R, 2014, IN C IND ENG ENG MAN, P1, DOI 10.1109/IEEM.2014.7058588
[4]  
Du Y, 2015, PROCEEDINGS 3RD IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION ACPR 2015, P579, DOI 10.1109/ACPR.2015.7486569
[5]  
Du Y, 2015, PROC CVPR IEEE, P1110, DOI 10.1109/CVPR.2015.7298714
[6]   Space-time representation of people based on 3D skeletal data: A review [J].
Han, Fei ;
Reily, Brian ;
Hoff, William ;
Zhang, Hao .
COMPUTER VISION AND IMAGE UNDERSTANDING, 2017, 158 :85-105
[7]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[8]  
Hu JF, 2015, PROC CVPR IEEE, P5344, DOI 10.1109/CVPR.2015.7299172
[9]   Differential Diagnosis of Atypical Hepatocellular Carcinoma in Contrast-Enhanced Ultrasound Using Spatio-Temporal Diagnostic Semantics [J].
Huang, Qinghua ;
Pan, Fengxin ;
Li, Wei ;
Yuan, Feiniu ;
Hu, Hangtong ;
Huang, Jinhua ;
Yu, Jie ;
Wang, Wei .
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2020, 24 (10) :2860-2869
[10]   Learning contrastive feature distribution model for interaction recognition [J].
Ji, Yanli ;
Cheng, Hong ;
Zheng, Yali ;
Li, Haoxin .
JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2015, 33 :340-349