Domain and View-Point Agnostic Hand Action Recognition

被引:22
作者
Sabater, Alberto [1 ]
Alonso, Inigo [1 ]
Montesano, Luis [1 ,2 ]
Murillo, Ana Cristina [1 ]
机构
[1] Univ Zaragoza, DIIS I3A, Zaragoza 50018, Spain
[2] Bitbrain Technol, Zaragoza 50008, Spain
关键词
Human and humanoid motion analysis and synthesis; gesture; posture and facial expressions; deep learning for visual perception; NETWORK;
D O I
10.1109/LRA.2021.3101822
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Hand action recognition is a special case of action recognition with applications in human-robot interaction, virtual reality or life-logging systems. Building action classifiers able to work for such heterogeneous action domains is very challenging. There are very subtle changes across different actions from a given application but also large variations across domains (e.g. virtual reality vs life-logging). This work introduces a novel skeleton-based hand motion representation model that tackles this problem. The framework we propose is agnostic to the application domain or camera recording view-point. When working on a single domain (intra-domain action classification) our approach performs better or similar to current state-of-the-art methods on well-known hand action recognition benchmarks. And, more importantly, when performing hand action recognition for action domains and camera perspectives which our approach has not been trained for (cross-domain action classification), our proposed framework achieves comparable performance to intra-domain state-of-the-art methods. These experiments show the robustness and generalization capabilities of our framework.
引用
收藏
页码:7823 / 7830
页数:8
相关论文
共 34 条
[1]  
Abbasi B, 2019, IEEE INT C INT ROBOT, P6756, DOI [10.1109/iros40897.2019.8968505, 10.1109/IROS40897.2019.8968505]
[2]  
[Anonymous], 2016, P 9 ISCA WORKSH SPEE
[3]  
Bai S., 2018, ARXIV PREPRINT ARXIV
[4]  
Bates T, 2017, IEEE INT C INT ROBOT, P3510, DOI 10.1109/IROS.2017.8206193
[5]   Learning a 3D Human Pose Distance Metric from Geometric Pose Descriptor [J].
Chen, Cheng ;
Zhuang, Yueting ;
Nie, Feiping ;
Yang, Yi ;
Wu, Fei ;
Xiao, Jun .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2011, 17 (11) :1676-1689
[6]  
Chen T, 2020, PR MACH LEARN RES, V119
[7]   MFA-Net: Motion Feature Augmented Network for Dynamic Hand Gesture Recognition from Skeletal Data [J].
Chen, Xinghao ;
Wang, Guijin ;
Guo, Hengkai ;
Zhang, Cairong ;
Wang, Hang ;
Zhang, Li .
SENSORS, 2019, 19 (02)
[8]  
De Smedt Q., 2017, 10 EUR WORKSH 3D OBJ, P1
[9]   Convolutional Two-Stream Network Fusion for Video Action Recognition [J].
Feichtenhofer, Christoph ;
Pinz, Axel ;
Zisserman, Andrew .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1933-1941
[10]   First-Person Hand Action Benchmark with RGB-D Videos and 3D Hand Pose Annotations [J].
Garcia-Hernando, Guillermo ;
Yuan, Shanxin ;
Baek, Seungryul ;
Kim, Tae-Kyun .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :409-419