ChoreoGraph: Music-conditioned Automatic Dance Choreography over a Style and Tempo Consistent Dynamic Graph

被引:15
作者
Au, Ho Yin [1 ]
Chen, Jie [1 ]
Jiang, Junkun [1 ]
Guo, Yike [1 ,2 ]
机构
[1] Hong Kong Baptist Univ, Dept Comp Sci, Hong Kong, Peoples R China
[2] Imperial Coll London, Data Sci Inst, London, England
来源
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022 | 2022年
关键词
3D Motion Synthesis; Cross-Modality Learning; Tempo Synchronization; Dynamic Motion Graph;
D O I
10.1145/3503161.3547797
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
To generate dance that temporally and aesthetically matches the music is a challenging problem, as the following factors need to be considered. First, the aesthetic styles and messages conveyed by the motion and music should be consistent. Second, the beats of the generated motion should be locally aligned to the musical features. And finally, basic choreomusical rules should be observed, and the motion generated should be diverse. To address these challenges, we propose ChoreoGraph, which choreographs high-quality dance motion for a given piece of music over a Dynamic Graph. A data-driven learning strategy is proposed to evaluate the aesthetic style and rhythmic connections between music and motion in a progressively learned cross-modality embedding space. The motion sequences will be beats-aligned based on the music segments and then incorporated as nodes of a Dynamic Motion Graph. Compatibility factors such as the style and tempo consistency, motion context connection, action completeness, and transition smoothness are comprehensively evaluated to determine the node transition in the graph. We demonstrate that our repertoire-based framework can generate motions with aesthetic consistency and robustly extensible in diversity. Both quantitative and qualitative experiment results show that our proposed model outperforms other baseline models.
引用
收藏
页码:3917 / 3925
页数:9
相关论文
共 34 条
[1]  
Aksan Emre, 2020, ARXIV200408692
[2]  
[Anonymous], 2015, ARXIV150800271
[3]  
[Anonymous], 2017, ARXIV170705363
[4]   Text2Gestures: A Transformer-Based Network for Generating Emotive Body Gestures for Virtual Agents [J].
Bhattacharya, Uttaran ;
Rewkowski, Nicholas ;
Banerjee, Abhishek ;
Guhan, Pooja ;
Bera, Aniket ;
Manocha, Dinesh .
2021 IEEE VIRTUAL REALITY AND 3D USER INTERFACES (VR), 2021, :160-169
[5]  
Bowden A, 2000, BUS WEEK, P10
[6]  
Chen K, 2021, ACM T GRAPHIC, V40, DOI [10.1145/3476576.3476724, 10.1145/3476576.3476641, 10.1145/3450626.3459932]
[7]   Action-Agnostic Human Pose Forecasting [J].
Chiu, Hsu-kuang ;
Adeli, Ehsan ;
Wang, Borui ;
Huang, De-An ;
Niebles, Juan Carlos .
2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2019, :1423-1432
[8]  
Cuturi M, 2017, PR MACH LEARN RES, V70
[9]   Bio-LSTM: A Biomechanically Inspired Recurrent Neural Network for 3-D Pedestrian Pose and Gait Prediction [J].
Du, Xiaoxiao ;
Vasudevan, Ram ;
Johnson-Roberson, Matthew .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (02) :1501-1508
[10]  
Heck R, 2007, I3D 2007: ACM SIGGRAPH SYMPOSIUM ON INTERACTIVE 3D GRAPHICS AND GAMES, PROCEEDINGS, P129