Automatic music emotion classification model for movie soundtrack subtitling based on neuroscientific premises

被引:7
作者
Lucia-Mulas, Maria Jose [1 ]
Revuelta-Sanz, Pablo [2 ]
Ruiz-Mezcua, Belen [1 ]
Gonzalez-Carrasco, Israel [1 ]
机构
[1] Univ Carlos III Madrid, Comp Sci Dept, Av Univ 20, Madrid 28915, Spain
[2] Univ Oviedo, C Luis Ortiz Berrocal S-N, Gijon 33203, Spain
关键词
Music emotion recognition; Automatic subtitling; Convolutional neural network; RECOGNITION;
D O I
10.1007/s10489-023-04967-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The ability of music to induce emotions has been arousing a lot of interest in recent years, especially due to the boom in music streaming platforms and the use of automatic music recommenders. Music Emotion Recognition approaches are based on combining multiple audio features extracted from digital audio samples and different machine learning techniques. In these approaches, neuroscience results on musical emotion perception are not considered. The main goal of this research is to facilitate the automatic subtitling of music. The authors approached the problem of automatic musical emotion detection in movie soundtracks considering these characteristics and using scientific musical databases, which have become a reference in neuroscience research. In the experiments, the Constant-Q-Transform spectrograms, the ones that best represent the relationships between musical tones from the point of view of human perception, are combined with Convolutional Neural Networks. Results show an efficient emotion classification model for 2-second musical audio fragments representative of intense basic feelings of happiness, sadness, and fear. Those emotions are the most interesting to be identified in the case of movie music captioning. The quality metrics have demonstrated that the results of the different models differ significantly and show no homogeneity. Finally, these results pave the way for an accessible and automatic captioning of music, which could automatically identify the emotional intent of the different segments of the movie soundtrack.
引用
收藏
页码:27096 / 27109
页数:14
相关论文
共 40 条
[21]  
Law Edith, 2009, ISMIR, DOI DOI 10.5281/ZENODO.1417647
[22]  
Li TLH, 2010, LECT NOTES ENG COMP, P546
[23]   Emotional classification of music using neural networks with the MediaEval dataset [J].
Medina, Yesid Ospitia ;
Beltran, Jose Ramon ;
Baldassarri, Sandra .
PERSONAL AND UBIQUITOUS COMPUTING, 2020, 26 (4) :1237-1249
[24]   The "Musical Emotional Bursts": a validated set of musical affect bursts to investigate auditory affective processing [J].
Paquette, Sebastien ;
Peretz, Isabelle ;
Belin, Pascal .
FRONTIERS IN PSYCHOLOGY, 2013, 4
[25]   How music alters a kiss: superior temporal gyrus controls fusiform-amygdalar effective connectivity [J].
Pehrs, Corinna ;
Deserno, Lorenz ;
Bakels, Jan-Hendrik ;
Schlochtermeier, Lorna H. ;
Kappelhoff, Hermann ;
Jacobs, Arthur M. ;
Fritz, Thomas Hans ;
Koelsch, Stefan ;
Kuchinke, Lars .
SOCIAL COGNITIVE AND AFFECTIVE NEUROSCIENCE, 2014, 9 (11) :1770-1778
[26]  
Peretz I., 2012, Handbook of Music and Emotion: Theory, P99
[27]  
Peretz I, 1998, Cogn, V68, P8
[28]   The circumplex model of affect: An integrative approach to affective neuroscience, cognitive development, and psychopathology [J].
Posner, J ;
Russell, JA ;
Peterson, BS .
DEVELOPMENT AND PSYCHOPATHOLOGY, 2005, 17 (03) :715-734
[29]   Design of optimal low-pass filter by a new Levy swallow swarm algorithm [J].
Sarangi, Shubhendu Kumar ;
Panda, Rutuparna ;
Abraham, Ajith .
SOFT COMPUTING, 2020, 24 (23) :18113-18128
[30]   FREQUENCY RATIOS AND THE PERCEPTION OF TONE PATTERNS [J].
SCHELLENBERG, EG ;
TREHUB, SE .
PSYCHONOMIC BULLETIN & REVIEW, 1994, 1 (02) :191-201