Towards noise robust acoustic insect detection: from the lab to the greenhouse

被引:5
作者
Branding, Jelto [1 ]
von Hoersten, Dieter [1 ]
Wegener, Jens Karl [1 ]
Boeckmann, Elias [2 ]
Hartung, Eberhard [3 ]
机构
[1] Julius Kuhn Inst JKI, Inst Applicat Tech Plant Protect, Messeweg 11-12, D-38104 Braunschweig, Germany
[2] Julius Kuhn Inst JKI, Inst Plant Protect Hort & Urban Green, Messeweg 11-12, D-38104 Braunschweig, Germany
[3] Christian Albrechts Univ Kiel, Inst Agr Proc Engn, Max Eyth Str 6, D-24118 Kiel, Germany
来源
KUNSTLICHE INTELLIGENZ | 2023年 / 37卷 / 2-4期
关键词
Insects; Audio; Acoustic; Identification; Recognition; Classification; Low-noise microphone; Microphone array; Anechoic box; Pest detection; Horticulture; Deep learning; WaveNet; Spectrogram; Raw audio; Neural beamforming; Background noise; Noise simulation; FREQUENCY;
D O I
10.1007/s13218-023-00812-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Successful and efficient pest management is key to sustainable horticultural food production. While greenhouses already allow digital monitoring and control of their climate parameters, a lack of digital pest sensors hinders the advent of digital pest management systems. To close the control loop, digital systems need to be enabled to directly assess the state of different insect populations in a greenhouse. The presented article investigates the feasibility of acoustic sensors for insect detection in greenhouses. The study is based on an extensive dataset of acoustic insect recordings made with an array of high-quality microphones under noise-shielded conditions. By mixing these noise-free laboratory recordings with environmental sounds recorded with the same equipment in a greenhouse, different signal-to-noise ratios (SNR) are simulated. To explore the possibilities of this unique and novel dataset, two deep-learning models are trained on this simulation data. A simple spectrogram-based model represents the baseline for a comparison with a model capable of processing multi-channel raw audio data. Making use of the unique possibility of the dataset, the models are pre-trained on clean data and fine-tuned on noisy data. Under lab conditions, results show that both models can make use of not just insect flight sounds but also the much quieter sounds of insect movements. First attempts under simulated real-world conditions showed the challenging nature of this task and the potential of spatial filtering. The analysis enabled by the proposed methods for training and evaluation provided valuable insights that should be considered for future work.
引用
收藏
页码:157 / 173
页数:17
相关论文
共 30 条
  • [21] Sainath TN, 2015, 16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, P1
  • [22] Sainath TN, 2016, INT CONF ACOUST SPEE, P5075, DOI 10.1109/ICASSP.2016.7472644
  • [23] Sainath TN, 2017, NEW ERA ROBUST SPEEC, P105, DOI [10.1007/978-3-319-64680-0_5, DOI 10.1007/978-3-319-64680-0_5]
  • [24] Shengyun Wei, 2020, Journal of Physics: Conference Series, V1453, DOI 10.1088/1742-6596/1453/1/012085
  • [25] Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
  • [26] SOTAVALTA O., 1947, ACTA ENT FENNICA, V4, P1
  • [27] Van Den Oord A., 2016, ARXIV160903499, V125, P2
  • [28] Virtanen P, 2020, NAT METHODS, V17, P261, DOI 10.1038/s41592-019-0686-2
  • [29] Weik MH, 2000, COMPUTER SCI COMMUNI, DOI [10.1007/1-4020-0613-6_19249, DOI 10.1007/1-4020-0613-6_19249]
  • [30] Two-Phase Object-Based Deep Learning for Multi-Temporal SAR Image Change Detection
    Zhang, Xinzheng
    Liu, Guo
    Zhang, Ce
    Atkinson, Peter M.
    Tan, Xiaoheng
    Jian, Xin
    Zhou, Xichuan
    Li, Yongming
    [J]. REMOTE SENSING, 2020, 12 (03)