Breathing Sound Segmentation and Detection Using Transfer Learning Techniques on an Attention-Based Encoder-Decoder Architecture

被引:1
作者
Hsiao, Chiu-Han [1 ]
Lin, Ting-Wei [2 ]
Lin, Chii-Wann [3 ]
Hsu, Fu-Shun [4 ,5 ]
Lin, Frank Yeong-Sung [2 ]
Chen, Chung-Wei [4 ,5 ]
Chung, Chi-Ming [6 ]
机构
[1] Acad Sinica, Res Ctr Informat Technol Innovat, Taipei 11529, Taiwan
[2] Natl Taiwan Univ, Informat Management Dept, Taipei 10617, Taiwan
[3] Natl Taiwan Univ, Dept Biomed Engn, Taipei 10617, Taiwan
[4] Natl Taiwan Univ, Grad Inst Biomed Elect & Bioinformat, New Taipei 22060, Taiwan
[5] Far Eastern Mem Hosp, Dept Crit Care Med, New Taipei 22060, Taiwan
[6] Natl Taiwan Univ, Comp Sci & Informat Engn Dept, Taipei 10617, Taiwan
来源
42ND ANNUAL INTERNATIONAL CONFERENCES OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY: ENABLING INNOVATIVE TECHNOLOGIES FOR GLOBAL HEALTHCARE EMBC'20 | 2020年
关键词
D O I
暂无
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
This paper focuses on the use of an attention-based encoder-decoder model for the task of breathing sound segmentation and detection. This study aims to accurately segment the inspiration and expiration of patients with pulmonary diseases using the proposed model. Spectrograms of the lung sound signals and labels for every time segment were used to train the model. The model would first encode the spectrogram and then detect inspiratory or expiratory sounds using the encoded image on an attention-based decoder. Physicians would be able to make a more precise diagnosis based on the more interpretable outputs with the assistance of the attention mechanism. The respiratory sounds used for training and testing were recorded from 22 participants using digital stethoscopes or anti-noising microphone sets. Experimental results showed a high 92.006% accuracy when applied 0.5 second time segments and ResNet101 as encoder. Consistent performance of the proposed method can be observed from ten-fold cross-validation experiments.
引用
收藏
页码:754 / 759
页数:6
相关论文
共 15 条
[1]  
Amoh Justice, 2015, 2015 IEEE Biomedical Circuits and Systems Conference (BioCAS), P1, DOI 10.1109/BioCAS.2015.7348395
[2]   Classification of lung sounds using convolutional neural networks [J].
Aykanat, Murat ;
Kilic, Ozkan ;
Kurt, Bahar ;
Saryal, Sevgi .
EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING, 2017,
[3]   Lung sounds classification using convolutional neural networks [J].
Bardou, Dalal ;
Zhang, Kun ;
Ahmad, Sayed Mohammad .
ARTIFICIAL INTELLIGENCE IN MEDICINE, 2018, 88 :58-69
[4]   Challenges in COVID-19: is pulmonary thromboembolism related to overall severity? [J].
Bellieni, Andrea ;
Intini, Enrica ;
Taddei, Eleonora ;
Baldi, Fabiana ;
Larosa, Luigi ;
Murri, Rita ;
Richeldi, Luca ;
Bernabei, Roberto ;
Cauda, Roberto ;
Landi, Francesco ;
Corbo, Giuseppe Maria ;
Fantoni, Massimo .
INFECTIOUS DISEASES, 2020, 52 (08) :585-589
[5]  
Chiu CC, 2018, 2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), P4774, DOI 10.1109/ICASSP.2018.8462105
[6]  
Clifford GD, 2016, COMPUT CARDIOL CONF, V43, P609
[7]  
Jacome Cristina, SENSORS, V20, P1214
[8]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[9]   Metrics for Polyphonic Sound Event Detection [J].
Mesaros, Annamaria ;
Heittola, Toni ;
Virtanen, Tuomas .
APPLIED SCIENCES-BASEL, 2016, 6 (06)
[10]  
Sutskever I, 2014, ADV NEUR IN, V27