Audio Features for Music Emotion Recognition: A Survey

被引:42
|
作者
Panda, Renato [1 ,2 ]
Malheiro, Ricardo [1 ,3 ]
Paiva, Rui Pedro [1 ]
机构
[1] Univ Coimbra, Ctr Informat & Syst, Dept Informat Engn, P-3030290 Coimbra, Portugal
[2] Polytech Inst Tomar, Ci2, P-2300313 Tomar, Portugal
[3] Miguel Torga Higher Inst, P-3000132 Coimbra, Portugal
关键词
Rhythm; Feature extraction; Emotion recognition; Psychology; Indexes; Machine learning; Affective computing; music emotion recognition; audio feature design; music information retrieval; PERCEPTION; EXPRESSION; PITCH; EXTRACTION; SPEECH; TIMBRE; REPRESENTATIONS; CLASSIFICATION; REGRESSION; RESPONSES;
D O I
10.1109/TAFFC.2020.3032373
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The design of meaningful audio features is a key need to advance the state-of-the-art in music emotion recognition (MER). This article presents a survey on the existing emotionally-relevant computational audio features, supported by the music psychology literature on the relations between eight musical dimensions (melody, harmony, rhythm, dynamics, tone color, expressivity, texture and form) and specific emotions. Based on this review, current gaps and needs are identified and strategies for future research on feature engineering for MER are proposed, namely ideas for computational audio features that capture elements of musical form, texture and expressivity that should be further researched. Previous MER surveys offered broad reviews, covering topics such as emotion paradigms, approaches for the collection of ground-truth data, types of MER problems and overviewing different MER systems. On the contrary, our approach is to offer a deep and specific review on one key MER problem: the design of emotionally-relevant audio features.
引用
收藏
页码:68 / 88
页数:21
相关论文
共 50 条
  • [21] EVALUATING MUSIC EMOTION RECOGNITION: LESSONS FROM MUSIC GENRE RECOGNITION?
    Sturm, Bob L.
    ELECTRONIC PROCEEDINGS OF THE 2013 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (ICMEW), 2013,
  • [22] Incorporating Interpersonal Synchronization Features for Automatic Emotion Recognition from Visual and Audio Data during Communication
    Quan, Jingyu
    Miyake, Yoshihiro
    Nozawa, Takayuki
    SENSORS, 2021, 21 (16)
  • [23] Audio-Visual Learning for Multimodal Emotion Recognition
    Fan, Siyu
    Jing, Jianan
    Wang, Chongwen
    SYMMETRY-BASEL, 2025, 17 (03):
  • [24] Review of data features-based music emotion recognition methods
    Yang, Xinyu
    Dong, Yizhuo
    Li, Juan
    MULTIMEDIA SYSTEMS, 2018, 24 (04) : 365 - 389
  • [25] EEG-Based BCI Emotion Recognition: A Survey
    Torres, Edgar P.
    Torres, Edgar A.
    Hernandez-Alvarez, Myriam
    Yoo, Sang Guun
    SENSORS, 2020, 20 (18) : 1 - 36
  • [26] Dimensional Music Emotion Recognition by Machine Learning
    Bai, Junjie
    Feng, Lixiao
    Peng, Jun
    Shi, Jinliang
    Luo, Kan
    Li, Zuojin
    Liao, Lu
    Wang, Yingxu
    INTERNATIONAL JOURNAL OF COGNITIVE INFORMATICS AND NATURAL INTELLIGENCE, 2016, 10 (04) : 74 - 89
  • [27] Music Theory-Inspired Acoustic Representation for Speech Emotion Recognition
    Li, Xingfeng
    Shi, Xiaohan
    Hu, Desheng
    Li, Yongwei
    Zhang, Qingchen
    Wang, Zhengxia
    Unoki, Masashi
    Akagi, Masato
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 2534 - 2547
  • [28] Ensemble Learning of Hybrid Acoustic Features for Speech Emotion Recognition
    Zvarevashe, Kudakwashe
    Olugbara, Oludayo
    ALGORITHMS, 2020, 13 (03)
  • [29] Pattern recognition and features selection for speech emotion recognition model using deep learning
    Jermsittiparsert, Kittisak
    Abdurrahman, Abdurrahman
    Siriattakul, Parinya
    Sundeeva, Ludmila A.
    Hashim, Wahidah
    Rahim, Robbi
    Maseleno, Andino
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2020, 23 (04) : 799 - 806
  • [30] A multi-genre model for music emotion recognition using linear regressors
    Griffiths, Darryl
    Cunningham, Stuart
    Weinel, Jonathan
    Picking, Richard
    JOURNAL OF NEW MUSIC RESEARCH, 2021, 50 (04) : 355 - 372