Windy events detection in big bioacoustics datasets using a pre-trained Convolutional Neural Network

被引:0
|
作者
Terranova, Francesca [1 ]
Betti, Lorenzo [2 ]
Ferrario, Valeria [1 ,3 ]
Friard, Olivier [1 ]
Ludynia, Katrin [4 ,5 ]
Petersen, Gavin Sean [4 ]
Mathevon, Nicolas [6 ,7 ,8 ]
Reby, David [6 ,7 ]
Favaro, Livio [1 ,9 ]
机构
[1] Univ Turin, Dept Life Sci & Syst Biol, Turin, Italy
[2] Cent European Univ, Dept Network & Data Sci, Vienna, Austria
[3] Chester Zoo, Caughall Rd, Chester, England
[4] Southern African Fdn Conservat Coastal Birds SANCC, Cape Town, South Africa
[5] Univ Western Cape, Dept Biodivers & Conservat Biol, Robert Sobukwe Rd, Bellville, South Africa
[6] Univ St Etienne, ENES Bioacoust Res Lab, CRNL, CNRS,Inserm, St Etienne, France
[7] Inst Univ France, Minist Higher Educ Res & Innovat, 1 rue Descartes, Paris, France
[8] PSL Univ, Ecole Prat Hautes Etud, CHArt lab, Paris, France
[9] Stn Zool Anton Dohrn, Naples, Italy
关键词
Bioacoustics; Deep learning; Ecoacoustics; Passive Acoustic Monitoring; Soundscape ecology; Wind-noise; ACOUSTIC INDEXES; TEMPERATE; KAPPA;
D O I
10.1016/j.scitotenv.2024.174868
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Passive Acoustic Monitoring (PAM), which involves using autonomous record units for studying wildlife behaviour and distribution, often requires handling big acoustic datasets collected over extended periods. While these data offer invaluable insights about wildlife, their analysis can present challenges in dealing with geophonic sources. A major issue in the process of detection of target sounds is represented by wind-induced noise. This can lead to false positive detections, i.e., energy peaks due to wind gusts misclassified as biological sounds, or false negative, i.e., the wind noise masks the presence of biological sounds. Acoustic data dominated by wind noise makes the analysis of vocal activity unreliable, thus compromising the detection of target sounds and, subsequently, the interpretation of the results. Our work introduces a straightforward approach for detecting recordings affected by windy events using a pre-trained convolutional neural network. This process facilitates identifying wind-compromised data. We consider this dataset pre-processing crucial for ensuring the reliable use of PAM data. We implemented this preprocessing by leveraging YAMNet, a deep learning model for sound classification tasks. We evaluated YAMNet as-is ability to detect wind-induced noise and tested its performance in a Transfer Learning scenario by using our annotated data from the Stony Point Penguin Colony in South Africa. While the classification of YAMNet as-is achieved a precision of 0.71, and recall of 0.66, those metrics strongly improved after the training on our annotated dataset, reaching a precision of 0.91, and recall of 0.92, corresponding to a relative increment of >28 %. Our study demonstrates the promising application of YAMNet in the bioacoustics and ecoacoustics fields, addressing the need for wind-noise-free acoustic data. We released an open-access code that, combined with the efficiency and peak performance of YAMNet, can be used on standard laptops for a broad user base.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Object detection and classification of butterflies using efficient CNN and pre-trained deep convolutional neural networks
    R. Faerie Mattins
    M. Vergin Raja Sarobin
    Azrina Abd Aziz
    S. Srivarshan
    Multimedia Tools and Applications, 2024, 83 : 48457 - 48482
  • [42] Object detection and classification of butterflies using efficient CNN and pre-trained deep convolutional neural networks
    Mattins, R. Faerie
    Sarobin, M. Vergin Raja
    Aziz, Azrina Abd
    Srivarshan, S.
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (16) : 48457 - 48482
  • [43] Analysis of Pre-trained Convolutional Neural Network Models in Diabetic Retinopathy Detection Through Retinal Fundus Images
    Escorcia-Gutierrez, Jose
    Cuello, Jose
    Barraza, Carlos
    Gamarra, Margarita
    Romero-Aroca, Pere
    Caicedo, Eduardo
    Valls, Aida
    Puig, Domenec
    COMPUTER INFORMATION SYSTEMS AND INDUSTRIAL MANAGEMENT (CISIM 2022), 2022, 13293 : 202 - 213
  • [44] Efficient Aspect Object Models Using Pre-trained Convolutional Neural Networks
    Wilkinson, Eric
    Takahashi, Takeshi
    2015 IEEE-RAS 15TH INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS (HUMANOIDS), 2015, : 284 - 289
  • [45] The performance comparison of pre-trained networks with the proposed lightweight convolutional neural network for disease detection in tomato leaves
    Ecemis, Irem Nur
    Ilhan, Hamza Osman
    JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY, 2023, 38 (02): : 693 - 705
  • [46] The Impact of Padding on Image Classification by Using Pre-trained Convolutional Neural Networks
    Tang, Hongxiang
    Ortis, Alessandro
    Battiato, Sebastiano
    IMAGE ANALYSIS AND PROCESSING - ICIAP 2019, PT II, 2019, 11752 : 337 - 344
  • [47] Classification of fresh water Zooplankton by pre-trained convolutional neural network in underwater microscopy
    Hong S.
    Mehdi S.R.
    Huang H.
    Shahani K.
    Zhang Y.
    Junaidullah
    Raza K.
    Khan M.A.
    International Journal of Advanced Computer Science and Applications, 2020, 11 (07): : 252 - 258
  • [48] Rapid seismic damage state prediction of the subway station structure using the pre-trained network and convolutional neural network
    Fan, Yifan
    Chen, Zhiyi
    Luo, Xiaowei
    SOIL DYNAMICS AND EARTHQUAKE ENGINEERING, 2024, 185
  • [49] EyeWeS: Weakly Supervised Pre-Trained Convolutional Neural Networks for Diabetic Retinopathy Detection
    Costa, Pedro
    Araujo, Teresa
    Aresta, Guilherme
    Galdran, Adrian
    Mendonca, Ana Maria
    Smailagic, Asim
    Campilho, Aurelio
    PROCEEDINGS OF MVA 2019 16TH INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS (MVA), 2019,
  • [50] Application of Pre-Trained Deep Convolutional Neural Networks for Coffee Beans Species Detection
    Unal, Yavuz
    Taspinar, Yavuz Selim
    Cinar, Ilkay
    Kursun, Ramazan
    Koklu, Murat
    FOOD ANALYTICAL METHODS, 2022, 15 (12) : 3232 - 3243