An Efficient Immersive Self-Training System for Hip-Hop Dance Performance with Automatic Evaluation Features

被引:2
作者
Esaki, Kazuhiro [1 ]
Nagao, Katashi [1 ]
机构
[1] Nagoya Univ, Grad Sch Informat, Nagoya 4648603, Japan
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 14期
关键词
virtual reality; dance training; automatic evaluation; deep learning; contrastive learning; MOTION CAPTURE; EMOTIONS;
D O I
10.3390/app14145981
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Featured Application Virtual Reality Simulation and Training for Dance Performance Improvement.Abstract As a significant form of physical expression, dance demands ongoing training for skill enhancement, particularly in expressiveness. However, such training often faces restrictions related to location and time. Moreover, the evaluation of dance performance tends to be subjective, which necessitates the development of effective training methods and objective evaluation techniques. In this research, we introduce a self-training system for dance that employs VR technology to create an immersive training environment that facilitates a comprehensive understanding of three-dimensional dance movements. Furthermore, the system incorporates markerless motion capture technology to accurately record dancers' movements in real time and translate them into the VR avatar. Additionally, the use of deep learning enables multi-perspective dance performance assessment, providing feedback to users to aid their repetitive practice. To enable deep learning-based dance evaluations, we established a dataset that incorporates data from beginner-level dances along with expert evaluations of those dances. This dataset was specifically curated for practitioners in a dance studio setting by using a total of four cameras to record dances. Expert annotations were obtained from various perspectives to provide a comprehensive evaluation. This study also proposes three unique automatic evaluation models. A comparative analysis of the models, particularly contrastive learning (and autoencoder)-based expression learning and a reference-guided model (where a model dancer's performance serves as a reference), revealed that the reference-guided model achieved superior accuracy. The proposed method was able to predict dance performance ratings with an accuracy of approximately +/- 1 point on a 10-point scale, compared to ratings by professional coaches. Our findings open up novel possibilities for future dance training and evaluation systems.
引用
收藏
页数:30
相关论文
共 62 条
[1]  
Ahir K, 2020, Augmented Human Research, V5, P7, DOI DOI 10.1007/S41133-019-0025-2
[2]   Human motions and emotions recognition inspired by LMA qualities [J].
Ajili, Insaf ;
Mallem, Malik ;
Didier, Jean-Yves .
VISUAL COMPUTER, 2019, 35 (10) :1411-1426
[3]  
Aristidou A., 2014, GCH 2014 - Eurographics Workshop on Graphics and Cultural Heritage, P55, DOI [DOI 10.2312/GCH.20141304, 10.2312/GCH.20141304]
[4]  
Bank D., 2020, arXiv
[5]  
Bazarevsky V, 2020, Arxiv, DOI [arXiv:2006.10204, DOI 10.48550/ARXIV.2006.10204]
[6]  
Bernstein R., 2015, Int. J. Comput. Electr. Autom. Control Inf. Eng., V9, P1574
[7]   DHP19: Dynamic Vision Sensor 3D Human Pose Dataset [J].
Calabrese, Enrico ;
Taverni, Gemma ;
Easthope, Christopher Awai ;
Skriabine, Sophie ;
Corradi, Federico ;
Longinotti, Luca ;
Eng, Kynan ;
Delbruck, Tobi .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, :1695-1704
[8]   A Virtual Reality Dance Training System Using Motion Capture Technology [J].
Chan, Jacky C. P. ;
Leung, Howard ;
Tang, Jeff K. T. ;
Komura, Taku .
IEEE TRANSACTIONS ON LEARNING TECHNOLOGIES, 2011, 4 (02) :187-195
[9]   Cross-View Tracking for Multi-Human 3D Pose Estimation at over 100 FPS [J].
Chen, Long ;
Ai, Haizhou ;
Chen, Rui ;
Zhuang, Zijie ;
Liu, Shuang .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :3276-3285
[10]  
Chen T, 2020, Arxiv, DOI [arXiv:2002.05709, DOI 10.48550/ARXIV.2002.05709]