Birdsong classification based on multi-feature fusion

被引:19
|
作者
Yan, Na [1 ]
Chen, Aibin [1 ,3 ]
Zhou, Guoxiong [1 ]
Zhang, Zhiqiang [2 ]
Liu, Xiangyong [4 ]
Wang, Jianwu [5 ]
Liu, Zhihua [1 ]
Chen, Wenjie [1 ]
机构
[1] Cent South Univ Forestry & Technol, Coll Comp & Informat Engn, Inst Artificial Intelligence Applicat, Changsha, Peoples R China
[2] Cent South Univ Forestry & Technol, Coll Forestry, Wildlife Conservat & Utilizat Lab, Changsha, Peoples R China
[3] Cent South Univ Forestry & Technol, Coll Life Sci & Technol, Hunan Prov Key Lab Urban Forest Ecol, Changsha, Peoples R China
[4] Hunan Zixing Artificial Intelligence Res Acad, Hunan Zixing, Peoples R China
[5] HuangFengQiao State Owned Forest Farm, Youxian Cty, Hunan, Peoples R China
关键词
Birdsong classification; Acoustic feature; Feature fusion; 3DCNN-LSTM; NEURAL-NETWORKS; RECOGNITION; SOUNDS; MFCC;
D O I
10.1007/s11042-021-11396-9
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The classification of birdsong has very important signification to monitor the bird population in the habitats. Aiming at the birdsong dataset with complex and diverse audio background, this paper attempts to introduce an acoustic feature for voice and music analysis: Chroma. It is spliced and fused with the commonly used birdsong features, Log-Mel Spectrogram (LM) and Mel Frequency Cepstrum Coefficient (MFCC), to enrich the representational capacity of single feature; At the same time, in view of the characteristic that birdsong has continuous and dynamic changes in time, a 3DCNN-LSTM combined model is proposed as a classifier to make the network more sensitive to the birdsong information that changes with time. In this paper, we selected four bird audio data from the Xeno-Canto website to evaluate how LM, MFCC and Chroma were fused to maximize the birdsong audio information. The experimental results show that the LM-MFCC-C feature combination achieves the best result of 97.9% mean average precision (mAP) in the experiment.
引用
收藏
页码:36529 / 36547
页数:19
相关论文
共 50 条
  • [31] Smoke Detection Based on Multi-feature Fusion
    Wu Dongmei
    Wang Nana
    Yan Hongmei
    2012 5TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING (CISP), 2012, : 220 - 223
  • [32] Subjectivity Detection Based on Multi-feature Fusion
    Tian, Weixin
    Sun, Shuifa
    Wang, Anhui
    2012 WORLD AUTOMATION CONGRESS (WAC), 2012,
  • [33] Knowledge tracing based on multi-feature fusion
    Yongkang Xiao
    Rong Xiao
    Ning Huang
    Yixin Hu
    Huan Li
    Bo Sun
    Neural Computing and Applications, 2023, 35 : 1819 - 1833
  • [34] Knowledge tracing based on multi-feature fusion
    Xiao, Yongkang
    Xiao, Rong
    Huang, Ning
    Hu, Yixin
    Li, Huan
    Sun, Bo
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (02): : 1819 - 1833
  • [35] Image retrieval based on multi-feature fusion
    Dong Wenfei
    Yu Shuchun
    Liu Songyu
    Zhang Zhiqiang
    Gu Wenbo
    2014 FOURTH INTERNATIONAL CONFERENCE ON INSTRUMENTATION AND MEASUREMENT, COMPUTER, COMMUNICATION AND CONTROL (IMCCC), 2014, : 240 - 243
  • [36] Multi-feature fusion dehazing based on CycleGAN
    Wang, Jingpin
    Ge, Yuan
    Zhao, Jie
    Han, Chao
    AI COMMUNICATIONS, 2024, 37 (04) : 619 - 635
  • [37] A lightweight multi-feature fusion structure for automatic modulation classification
    Li, Zhigang
    Zhang, Wentao
    Wang, Yutong
    Li, Shujie
    Sun, Xiaochuan
    PHYSICAL COMMUNICATION, 2023, 61
  • [38] Multi-feature structure fusion of contours for unsupervised shape classification
    Lin, Guangfeng
    Zhu, Hong
    Rang, Xiaobing
    Fan, Caixia
    Zhang, Erhu
    PATTERN RECOGNITION LETTERS, 2013, 34 (11) : 1286 - 1290
  • [39] Multi-Feature Extraction and Fusion for the Underwater Moving Targets Classification
    Yang Juan
    Xu Feng
    Wei Zhiheng
    Liu Jia
    An Xudong
    SENSORS, MEASUREMENT AND INTELLIGENT MATERIALS, PTS 1-4, 2013, 303-306 : 1357 - 1360
  • [40] EEG FEATURE EXTRACTION AND RECOGNITION BASED ON MULTI-FEATURE FUSION
    Sun, Jian
    Wu, Quanyu
    Gao, Nan
    Pan, Lingjiao
    Tao, Weige
    BIOMEDICAL ENGINEERING-APPLICATIONS BASIS COMMUNICATIONS, 2024, 36 (06):