Machine Learning and Deep Learning in Music Emotion Recognition: A Comprehensive Survey

被引:0
作者
Dutta, Jumpi [1 ]
Chanda, Dipankar [1 ]
机构
[1] Assam Engn Coll, Dept Elect Engn, Gauhati, Assam, India
关键词
Music emotion recognition; Feature extraction; Machine learning; Deep learning; Human-computer-interaction; CLASSIFICATION; MODEL; EXPRESSION;
D O I
10.33889/IJMEMS.2025.10.4.047
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Music can express and influence a wide range of emotional states and feelings in humans. The development of a system for recognizing emotions based on music analysis has generated significant interest among academic and industrial communities due to its applications in various fields such as human-machine interaction, music recommendation systems, music therapy, and so on. Music emotion recognition (MER) is the process of analysing and classifying the affective states conveyed by a piece of music. A survey of existing work on emotional music processing is indeed very helpful for carrying out further research in the field of music emotion recognition. Due to the importance of emotion recognition in Music Information Retrieval (MIR) research, a comprehensive survey is provided in this paper with a detailed study of emotion models, features, and various music databases. This paper emphasizes the machine learning and deep learning approaches used for MER to extract emotions from music. Finally, the paper is summarized with some possible future research directions.
引用
收藏
页码:977 / 999
页数:23
相关论文
共 100 条
[1]  
Agarwal G, 2023, International Journal of Reconfigurable and Embedded Systems (IJRES), V12, P240, DOI [10.11591/ijres.v12.i2.pp240-247, 10.11591/ijres.v12.i2, DOI 10.11591/IJRES.V12.I2]
[2]   A systematic survey on multimodal emotion recognition using learning algorithms [J].
Ahmed, Naveed ;
Al Aghbari, Zaher ;
Girija, Shini .
INTELLIGENT SYSTEMS WITH APPLICATIONS, 2023, 17
[3]   Developing a benchmark for emotional analysis of music [J].
Aljanaki, Anna ;
Yang, Yi-Hsuan ;
Soleymani, Mohammad .
PLOS ONE, 2017, 12 (03)
[4]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[5]  
Barthet M, 2013, LECT NOTES COMPUT SC, V7900, P228, DOI 10.1007/978-3-642-41248-6_13
[6]  
Bhakre SK, 2016, 2016 INTERNATIONAL CONFERENCE ON ADVANCES IN COMPUTING, COMMUNICATIONS AND INFORMATICS (ICACCI), P2363, DOI 10.1109/ICACCI.2016.7732408
[7]  
Bogdanov D., 2022, P 23 INT SOC MUS INF, P05, DOI DOI 10.5281/ZENODO.7316746
[8]  
Cabrera D., 1999, Australian Acoustical Society Conference, V24, P47, DOI [10.1002/asi, DOI 10.1002/ASI]
[9]  
Chandrappa S, 2023, SN Computer Science, V4, DOI [10.1007/s42979-023-01812-2, 10.1007/s42979-023-01812-2, DOI 10.1007/S42979-023-01812-2]
[10]   The Emotional Characteristics and Timbre of Nonsustaining Instrument Sounds [J].
Chau, Chuck-Jee ;
Wu, Bin ;
Horner, Andrew .
JOURNAL OF THE AUDIO ENGINEERING SOCIETY, 2015, 63 (04) :228-244