Soul Dancer: Emotion-Based Human Action Generation

被引:5
作者
Hou, Yuxin [1 ]
Yao, Hongxun [1 ]
Sun, Xiaoshuai [1 ]
Li, Haoran [1 ]
机构
[1] Harbin Inst Technol, Dazhi St West 92, Harbin, Heilongjiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Human emotion; body language; pose sequence; action generation; EXPRESSION ANALYSIS; SENTIMENT ANALYSIS; RECOGNITION;
D O I
10.1145/3340463
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Body language is one of the most common ways of expressing human emotion. In this article, we make the first attempt to generate an action video with a specific emotion from a single person image. The goal of the emotion-based action generation task (EBAG) is to generate action videos expressing a specific type of emotion given a single reference image with a full human body. We divide the task into two parts and propose a two-stage framework to generate action videos with specified emotions. At the first stage, we propose an emotion-based pose sequence generation approach (EPOSE-GAN) for translating the emotion to a pose sequence. At the second stage, we generate the target video frames according to the three inputs including the source pose and the target pose as the motion information and the source image as the appearance reference by using conditional GAN model with an online training strategy. Our framework produces the pose sequence and transforms the action independently, which highlights the fundamental role that the high-level pose feature plays in generating action video with a specific emotion. The proposed method has been evaluated on the "Soul Dancer" dataset which is built for action emotion analysis and generation. The experimental results demonstrate that our framework can effectively solve the emotion-based action generation task. However, the gap in the details of the appearance between the generated action video and the real-world video still exists, which indicates that the emotion-based action generation task has great research potential.
引用
收藏
页数:19
相关论文
共 72 条
[1]  
Aman S, 2007, LECT NOTES ARTIF INT, V4629, P196
[2]  
[Anonymous], ARXIV181011794CS
[3]  
[Anonymous], P EUR C COMP VIS ECC
[4]  
[Anonymous], P 3 INT C LEARNING R
[5]  
[Anonymous], 2016, P IEEE C COMP VIS PA
[6]  
[Anonymous], 2014, ADV NEURAL INFORM PR
[7]  
[Anonymous], ARXIV171201955CS
[8]  
[Anonymous], 2016, LECT NOTES COMPUT SC, DOI DOI 10.1007/978-3-319-46484-8_29
[9]  
[Anonymous], 2017, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2017.632
[10]  
[Anonymous], 2018, CORR