Music genre classification based on auditory image, spectral and acoustic features

被引:0
|
作者
Xin Cai
Hongjuan Zhang
机构
[1] Shanghai University,Department of Mathematics
来源
Multimedia Systems | 2022年 / 28卷
关键词
Music genre classification; Auditory image feature; Spectral feature; Acoustic feature; Feature fusion;
D O I
暂无
中图分类号
学科分类号
摘要
Music genre is one of the conventional ways to describe music content, and also is one of the important labels of music information retrieval. Therefore, the effective and precise music genre classification method becomes an urgent need for realizing automatic organization of large music archives. Inspired by the fact that humans have a better automatic recognizing music genre ability, which may attribute to our auditory system, even for the participants with little musical literacy. In this paper, a novel classification framework incorporating the auditory image feature with traditional acoustic features and spectral feature is proposed to improve the classification accuracy. In detail, auditory image feature is extracted based on the auditory image model which simulates the auditory system of the human ear and has also been successfully used in other fields apart from music genre classification to our best knowledge. Moreover, the logarithmic frequency spectrogram rather than linear is adopted to extract the spectral feature to capture the information about the low-frequency part adequately. These above two features and the traditional acoustic feature are evaluated, compared, respectively, and fused finally based on the GTZAN, GTZAN-NEW, ISMIR2004 and Homburg datasets. Experimental results show that the proposed method owns the higher classification accuracy and the better stability than many state-of-the-art classification methods.
引用
收藏
页码:779 / 791
页数:12
相关论文
共 50 条
  • [1] Music genre classification based on auditory image, spectral and acoustic features
    Cai, Xin
    Zhang, Hongjuan
    MULTIMEDIA SYSTEMS, 2022, 28 (03) : 779 - 791
  • [2] Automatic Music Genre Classification Based on Modulation Spectral Analysis of Spectral and Cepstral Features
    Lee, Chang-Hsing
    Shih, Jau-Ling
    Yu, Kun-Ming
    Lin, Hwai-San
    IEEE TRANSACTIONS ON MULTIMEDIA, 2009, 11 (04) : 670 - 682
  • [3] Combining Acoustic and Multilevel Visual Features for Music Genre Classification
    Wu, Ming-Ju
    Jang, Jyh-Shing R.
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2015, 12 (01) : 1 - 17
  • [4] A Short Survey and Comparison of CNN-Based Music Genre Classification Using Multiple Spectral Features
    Seo, Wangduk
    Cho, Sung-Hyun
    Teisseyre, Pawe
    Lee, Jaesung
    IEEE ACCESS, 2024, 12 : 245 - 257
  • [5] Music Features based on Hu Moments for Genre Classification
    Lopes, Renia
    Chapaneri, Santosh
    Jayaswal, Deepak
    2017 2ND INTERNATIONAL CONFERENCE ON COMMUNICATION SYSTEMS, COMPUTING AND IT APPLICATIONS (CSCITA), 2017, : 22 - 27
  • [6] Music Genre Classification Based on Chroma Features and Deep Learning
    Shi, Leisi
    Li, Chen
    Tian, Lihua
    2019 TENTH INTERNATIONAL CONFERENCE ON INTELLIGENT CONTROL AND INFORMATION PROCESSING (ICICIP), 2019, : 81 - 86
  • [7] Robust handcrafted features for music genre classification
    Victor Hugo da Silva Muniz
    João Baptista de Oliveira e Souza Filho
    Neural Computing and Applications, 2023, 35 : 9335 - 9348
  • [8] Robust handcrafted features for music genre classification
    Muniz, Victor Hugo da Silva
    de Oliveira e Souza Filho, Joao Baptista
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (13) : 9335 - 9348
  • [9] COMPARISON OF DIFFERENT REPRESENTATIONS BASED ON NONLINEAR FEATURES FOR MUSIC GENRE CLASSIFICATION
    Zlatintsi, Athanasia
    Maragos, Petros
    2014 PROCEEDINGS OF THE 22ND EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2014, : 1547 - 1551
  • [10] Exploring Textural Features for Automatic Music Genre Classification
    Agera, Nelson
    Chapaneri, Santosh
    Jayaswal, Deepak
    1ST INTERNATIONAL CONFERENCE ON COMPUTING COMMUNICATION CONTROL AND AUTOMATION ICCUBEA 2015, 2015, : 822 - 826