Music Style Classification Algorithm Based on Music Feature Extraction and Deep Neural Network

被引:28
作者
Zhang, Kedong [1 ]
机构
[1] Xian Conservatory Mus, Dept Vocal Mus, Xian 710061, Shaanxi, Peoples R China
关键词
PERCEIVED EMOTIONS;
D O I
10.1155/2021/9298654
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The music style classification technology can add style tags to music based on the content. When it comes to researching and implementing aspects like efficient organization, recruitment, and music resource recommendations, it is critical. Traditional music style classification methods use a wide range of acoustic characteristics. The design of characteristics necessitates musical knowledge and the characteristics of various classification tasks are not always consistent. The rapid development of neural networks and big data technology has provided a new way to better solve the problem of music-style classification. This paper proposes a novel method based on music extraction and deep neural networks to address the problem of low accuracy in traditional methods. The music style classification algorithm extracts two types of features as classification characteristics for music styles: timbre and melody features. Because the classification method based on a convolutional neural network ignores the audio's timing. As a result, we proposed a music classification module based on the one-dimensional convolution of a recurring neuronal network, which we combined with single-dimensional convolution and a two-way, recurrent neural network. To better represent the music style properties, different weights are applied to the output. The GTZAN data set was also subjected to comparison and ablation experiments. The test results outperformed a number of other well-known methods, and the rating performance was competitive.
引用
收藏
页数:7
相关论文
共 29 条
[1]   Mel-Frequency Cepstral Coefficient Features Based on Standard Deviation and Principal Component Analysis for Language Identification Systems [J].
Albadr, Musatafa Abbas Abbood ;
Tiun, Sabrina ;
Ayob, Masri ;
Mohammed, Manal ;
AL-Dhief, Fahad Taha .
COGNITIVE COMPUTATION, 2021, 13 (05) :1136-1153
[2]  
[Anonymous], 2018, ARXIV180401149
[3]   RELATIONSHIPS BETWEEN PERCEIVED EMOTIONS IN MUSIC AND MUSIC-INDUCED MOVEMENT [J].
Burger, Birgitta ;
Saarikallio, Suvi ;
Luck, Geoff ;
Thompson, Marc R. ;
Toiviainen, Petri .
MUSIC PERCEPTION, 2013, 30 (05) :517-533
[4]   An Adaptive Machine Learning Method Based on Finite Element Analysis for Ultra Low-k Chip Package Design [J].
Chu, Weishen ;
Ho, Paul S. ;
Li, Wei .
IEEE TRANSACTIONS ON COMPONENTS PACKAGING AND MANUFACTURING TECHNOLOGY, 2021, 11 (09) :1435-1441
[5]   Digital Twins The Convergence of Multimedia Technologies [J].
El Saddik, Abdulmotaleb .
IEEE MULTIMEDIA, 2018, 25 (02) :87-92
[6]  
Fleischer Rasmus., 2017, CULTURE UNBOUND J CU, V9, P146
[7]   Music Feature Classification Based on Recurrent Neural Networks with Channel Attention Mechanism [J].
Gan, Jie .
MOBILE INFORMATION SYSTEMS, 2021, 2021
[8]  
Guimaraes P., 2020, P 4 WIDENING NATURAL, P61
[9]  
Ick Kang Sang, 2017, [Journal of Rehabilitation Welfare Engineering & Assistive Technology, 재활복지공학회논문지], V11, P143, DOI 10.21288/resko.2017.11.2.143
[10]   A Novel Negative-Transfer-Resistant Fuzzy Clustering Model With a Shared Cross-Domain Transfer Latent Space and its Application to Brain CT Image Segmentation [J].
Jiang, Yizhang ;
Gu, Xiaoqing ;
Wu, Dongrui ;
Hang, Wenlong ;
Xue, Jing ;
Qiu, Shi ;
Lin, Chin-Teng .
IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2021, 18 (01) :40-52