FILTWAM and Voice Emotion Recognition

被引:1
作者
Bahreini, Kiavash [1 ]
Nadolski, Rob [1 ]
Westera, Wim [1 ]
机构
[1] Open Univ Netherlands, Ctr Learning Sci & Technol CELSTEC, NL-6419 AT Heerlen, Netherlands
来源
GAMES AND LEARNING ALLIANCE | 2014年 / 8605卷
关键词
Game-based learning; Human-computer interaction; Multimodal emotion recognition; Real-time voice emotion recognition; Microphone; AUDIO; GAMES;
D O I
10.1007/978-3-319-12157-4_10
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper introduces the voice emotion recognition part of our framework for improving learning through webcams and microphones (FILTWAM). This framework enables multimodal emotion recognition of learners during game-based learning. The main goal of this study is to validate the use of microphone data for a real-time and adequate interpretation of vocal expressions into emotional states were the software is calibrated with end users. FILTWAM already incorporates a valid face emotion recognition module and is extended with a voice emotion recognition module. This extension aims to provide relevant and timely feedback based upon learner's vocal intonations. The feedback is expected to enhance learner's awareness of his or her own behavior. Six test persons received the same computer-based tasks in which they were requested to mimic specific vocal expressions. Each test person mimicked 82 emotions, which led to a dataset of 492 emotions. All sessions were recorded on video. An overall accuracy of our software based on the requested emotions and the recognized emotions is a pretty good 74.6 % for the emotions happy and neutral emotions; but will be improved for the lower values of an extended set of emotions. In contrast with existing software our solution allows to continuously and unobtrusively monitor learners' intonations and convert these intonations into emotional states. This paves the way for enhancing the quality and efficacy of game-based learning by including the learner's emotional states, and links these to pedagogical scaffolding.
引用
收藏
页码:116 / 129
页数:14
相关论文
共 34 条
[31]  
Song ML, 2004, PROC CVPR IEEE, P1020
[32]  
Subramanian R., 2010, PUTTING PIECES TOGET
[33]  
Wagner J., 2011, P INTERSPEECH FLOR I
[34]   A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions [J].
Zeng, Zhihong ;
Pantic, Maja ;
Roisman, Glenn I. ;
Huang, Thomas S. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2009, 31 (01) :39-58