Deep neural network based environment sound classification and its implementation on hearing aid app

被引:20
作者
Fan, Xiaoqian [1 ]
Sun, Tianyi [1 ]
Chen, Wenzhi [1 ]
Fan, Quanfang [2 ]
机构
[1] Zhejiang Univ, Sch Comp Sci & Technol, 38 Zheda Rd, Hangzhou 310007, Peoples R China
[2] Hangzhou Youting Technol Co Ltd, 220,East Zone,Bldg A,525 Xixi Rd, Hangzhou 310007, Peoples R China
关键词
Sound environment classification; Superimposed audio blocks(SAB); Deep neural networks; Feature extraction; Hearing aids; RECOGNITION;
D O I
10.1016/j.measurement.2020.107790
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
In general, a hearing aid app is very useful for the persons having either partial or complete inability to hear. At present, there is no special provision available in the hearing aid app for the classification of different environmental sounds. This paper proposes an algorithm for environmental sound classification based on Superimposed Audio Blocks using Deep Neural Networks (SAB - DNN) and also to implement it on the hearing aid app. The system can recognize five kinds of different sound fields automatically: bus, subway, street, indoor, car. In this system, 512 sampling points are taken as an audio frame and several audio frames are stacked up into an Audio Block (AB). when 7 audio frames are stacked up into an Audio Block (AB), the accuracy rate of sound environment classification using (AB - DNN) tends to be the best (96.18%). Based on this, the experiment integrates multiple Audio Block (AB) into an audio unit called Superimposed Audio Blocks(SAB) and classify it using DNN. Optimally, 30 sound blocks are integrated into a SAB which results in the classification accuracy up to 98.8%. As far as we know, it is the first time on the hearing aid app to implement an improved Deep Neural Network (DNN) based classification system and superposition of multi-audio frames and blocks. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页数:7
相关论文
共 29 条
[1]   Feature selection for sound classification in hearing aids through restricted search driven by genetic algorithms [J].
Alexandre, Enrique ;
Cuadra, Lucas ;
Rosa, Manuel ;
Lopez-Ferreras, Francisco .
IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2007, 15 (08) :2249-2256
[2]  
[Anonymous], 2020, BIORXIV
[3]  
[Anonymous], CROSSTASK LEARNING A
[4]   The bag-of-frames approach to audio pattern recognition: A sufficient model for urban soundscapes but not for polyphonic music [J].
Aucouturier, Jean-Julien ;
Defreville, Boris ;
Pachet, Francois .
JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2007, 122 (02) :881-891
[5]   Automatic Scene Recognition through Acoustic Classification for Behavioral Robotics [J].
Aziz, Sumair ;
Awais, Muhammad ;
Akram, Tallha ;
Khan, Umar ;
Alhussein, Musaed ;
Aurangzeb, Khursheed .
ELECTRONICS, 2019, 8 (05)
[6]   Acoustic Scene Classification [J].
Barchiesi, Daniele ;
Giannoulis, Dimitrios ;
Stowell, Dan ;
Plumbley, Mark D. .
IEEE SIGNAL PROCESSING MAGAZINE, 2015, 32 (03) :16-34
[7]   Sound classification in hearing aids inspired by auditory scene analysis [J].
Büchler, M ;
Allegro, S ;
Launer, S ;
Dillier, N .
EURASIP JOURNAL ON APPLIED SIGNAL PROCESSING, 2005, 2005 (18) :2991-3002
[8]  
Chachada S, 2013, ASIAPAC SIGN INFO PR
[9]  
Chaudhuri S, 2013, INT CONF ACOUST SPEE, P833, DOI 10.1109/ICASSP.2013.6637765
[10]   Environmental Sound Recognition With Time-Frequency Audio Features [J].
Chu, Selina ;
Narayanan, Shrikanth ;
Kuo, C. -C. Jay .
IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2009, 17 (06) :1142-1158