The Jena Audiovisual Stimuli of Morphed Emotional Pseudospeech (JAVMEPS): A database for emotional auditory-only, visual-only, and congruent and incongruent audiovisual voice and dynamic face stimuli with varying voice intensities

被引:0
作者
von Eiff, Celina I. [1 ,2 ,3 ,4 ]
Kauk, Julian [1 ]
Schweinberger, Stefan R. [1 ,2 ,3 ,4 ]
机构
[1] Friedrich Schiller Univ Jena, Inst Psychol, Dept Gen Psychol & Cognit Neuroscience, D-307743JEN Am Steiger, Germany
[2] Friedrich Schiller Univ Jena, Inst Psychol, Voice Res Unit, D-107743JEN Leutragraben, Germany
[3] DFG SPP 2392 Visual Commun ViCom, Frankfurt, Germany
[4] Jena Univ Hosp, Jena, Germany
关键词
Emotion; Audiovisual integration; Voice morphing; Stimulus database; Adaptive testing; Emotion induction; Cochlear implant; FACIAL EXPRESSIONS; NONVERBAL DIALECTS; PERCEPTION; RECOGNITION; INTEGRATION; SPECTRUM; SET; REPRESENTATION; APERIODICITY; INFORMATION;
D O I
10.3758/s13428-023-02249-4
中图分类号
B841 [心理学研究方法];
学科分类号
040201 ;
摘要
We describe JAVMEPS, an audiovisual (AV) database for emotional voice and dynamic face stimuli, with voices varying in emotional intensity. JAVMEPS includes 2256 stimulus files comprising (A) recordings of 12 speakers, speaking four bisyllabic pseudowords with six naturalistic induced basic emotions plus neutral, in auditory-only, visual-only, and congruent AV conditions. It furthermore comprises (B) caricatures (140%), original voices (100%), and anti-caricatures (60%) for happy, fearful, angry, sad, disgusted, and surprised voices for eight speakers and two pseudowords. Crucially, JAVMEPS contains (C) precisely time-synchronized congruent and incongruent AV (and corresponding auditory-only) stimuli with two emotions (anger, surprise), (C1) with original intensity (ten speakers, four pseudowords), (C2) and with graded AV congruence (implemented via five voice morph levels, from caricatures to anti-caricatures; eight speakers, two pseudowords). We collected classification data for Stimulus Set A from 22 normal-hearing listeners and four cochlear implant users, for two pseudowords, in auditory-only, visual-only, and AV conditions. Normal-hearing individuals showed good classification performance (M-corrAV = .59 to .92), with classification rates in the auditory-only condition >= .38 correct (surprise: .67, anger: .51). Despite compromised vocal emotion perception, CI users performed above chance levels of .14 for auditory-only stimuli, with best rates for surprise (.31) and anger (.30). We anticipate JAVMEPS to become a useful open resource for researchers into auditory emotion perception, especially when adaptive testing or calibration of task difficulty is desirable. With its time-synchronized congruent and incongruent stimuli, JAVMEPS can also contribute to filling a gap in research regarding dynamic audiovisual integration of emotion perception via behavioral or neurophysiological recordings.
引用
收藏
页码:5103 / 5115
页数:13
相关论文
共 101 条
[51]   Presentation and validation of the Radboud Faces Database [J].
Langner, Oliver ;
Dotsch, Ron ;
Bijlstra, Gijsbert ;
Wigboldus, Daniel H. J. ;
Hawk, Skyler T. ;
van Knippenberg, Ad .
COGNITION & EMOTION, 2010, 24 (08) :1377-1388
[52]   Expression of affect in spontaneous speech: Acoustic correlates and automatic detection of irritation and resignation [J].
Laukka, Petri ;
Neiberg, Daniel ;
Forsell, Mimmi ;
Karlsson, Inger ;
Elenius, Kjell .
COMPUTER SPEECH AND LANGUAGE, 2011, 25 (01) :84-104
[53]   The Ryerson Audio-Visual Database of Emotional Speech and Song (RAVDESS): A dynamic, multimodal set of facial and vocal expressions in North American English [J].
Livingstone, Steven R. ;
Russo, Frank A. .
PLOS ONE, 2018, 13 (05)
[54]  
Lundqvist D., 1998, The Karolinska Directed Emotional Faces-KDEF. CD ROM from Department of Clinical Neuroscience, V91, P2
[55]  
Matsumoto D., 1988, Japanese and Caucasian Facial Expressions of Emotion (JACFEE) and Neutral Faces (JACNeuF)
[56]   A NEW SERIES OF SLIDES DEPICTING FACIAL EXPRESSIONS OF AFFECT - A COMPARISON WITH THE PICTURES OF FACIAL AFFECT SERIES [J].
MAZURSKI, EJ ;
BOND, NW .
AUSTRALIAN JOURNAL OF PSYCHOLOGY, 1993, 45 (01) :41-47
[57]   Temporal constraints on the McGurk effect [J].
Munhall, KG ;
Gribble, P ;
Sacco, L ;
Ward, M .
PERCEPTION & PSYCHOPHYSICS, 1996, 58 (03) :351-362
[58]   Dynamic Facial Expressions Allow Differentiation of Displays Intended to Convey Positive and Hubristic Pride [J].
Nelson, Nicole L. ;
Russell, James A. .
EMOTION, 2014, 14 (05) :857-864
[59]   Musical and vocal emotion perception for cochlear implants users [J].
Paquette, S. ;
Ahmed, G. D. ;
Goffi-Gomez, M. V. ;
Hoshino, A. C. H. ;
Peretz, I. ;
Lehmann, A. .
HEARING RESEARCH, 2018, 370 :272-282
[60]   Recognizing Emotions in a Foreign Language [J].
Pell, Marc D. ;
Monetta, Laura ;
Paulmann, Silke ;
Kotz, Sonja A. .
JOURNAL OF NONVERBAL BEHAVIOR, 2009, 33 (02) :107-120