Prediction of musical affect using a combination of acoustic structural cues

被引:43
作者
Leman, M [1 ]
Vermeulen, V [1 ]
De Voogdt, L [1 ]
Moelants, D [1 ]
Lesaffre, M [1 ]
机构
[1] Univ Ghent, Dept Musicol, IPEM, B-9000 Ghent, Belgium
关键词
D O I
10.1080/09298210500123978
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
This study explores whether musical affect attribution can be predicted by a linear combination of acoustical structural cues. To that aim, a database of sixty musical audio excerpts was compiled and analyzed at three levels: judgments of affective content by subjects; judgments of structural content by musicological experts (i.e., "manual structural cues"), and extraction of structural content by an auditory-based computer algorithm (called: acoustical structural cues). In Study 1, an affect space was constructed with Valence (gay-sad), Activity (tender-bold) and Interest (exciting-boring) as the main dimensions, using the responses of a hundred subjects. In Study 11 manual and acoustical structural cues were analyzed and compared. Manual structural cues such as loudness and articulation could be accounted for in terms of a combination of acoustical structural cues. In Study 111, the subjective responses of eight individual subjects were analyzed using the affect space obtained in Study 1, and modeled in terms of the structural cues obtained in Study 11, using linear regression modeling. This worked better for the Activity dimension than for the Valence dimension, while the Interest dimension could not be accounted for. Overall, manual structural cues worked better than acoustical structural cues. In a final assessment study, a selected set of acoustical structural cues was used for building prediction models. The results indicate that musical affect attribution can partly be predicted using a combination of acoustical structural cues. Future research may focus on non-linear approaches, elaboration of dataset and subjects, and refinement of acoustical structural cue extraction.
引用
收藏
页码:39 / 67
页数:29
相关论文
共 73 条
[21]   AUDITORY MODELING AND SELF-ORGANIZING NEURAL NETWORKS FOR TIMBRE CLASSIFICATION [J].
COSI, P ;
DEPOLI, G ;
LAUZZANA, G .
JOURNAL OF NEW MUSIC RESEARCH, 1994, 23 (01) :71-98
[22]  
Daniel P, 1997, ACUSTICA, V83, P113
[23]  
DANNENBERG R, 2002, INT S MUS INF RETR 2
[24]  
DEMULDER T, 2004, IEEE INT C AC SPEECH
[25]   Sonological models for timbre characterization [J].
DePoli, G ;
Prandoni, P .
JOURNAL OF NEW MUSIC RESEARCH, 1997, 26 (02) :170-197
[26]   The role of peripheral feedback in emotional experience with music [J].
Dibben, N .
MUSIC PERCEPTION, 2004, 22 (01) :79-115
[27]   A dynamical systems interpretation of a dimensional model of emotion [J].
Faith, M ;
Thayer, JF .
SCANDINAVIAN JOURNAL OF PSYCHOLOGY, 2001, 42 (02) :121-133
[28]  
Gabrielsson A, 2003, SER AFFECTIVE SCI, P503
[29]   Mode and tempo relative contributions to "happy-sad" judgements in equitone melodies [J].
Gagnon, L ;
Peretz, I .
COGNITION & EMOTION, 2003, 17 (01) :25-40
[30]   Affective and physiological responses to environmental noises and music [J].
Gomez, P ;
Danuser, B .
INTERNATIONAL JOURNAL OF PSYCHOPHYSIOLOGY, 2004, 53 (02) :91-103