The Phonetic Footprint of Covid-19?

被引:4
作者
Klumpp, P. [1 ]
Bocklet, T. [2 ]
Arias-Vergara, T. [1 ,3 ,4 ]
Vasquez-Correa, J. C. [1 ,3 ]
Perez-Toro, P. A. [1 ,3 ]
Bayerl, S. P. [2 ]
Orozco-Arroyave, J. R. [1 ,3 ]
Noeth, Elmar [1 ]
机构
[1] Friedrich Alexander Univ, Pattern Recognit Lab, Erlangen, Germany
[2] TH Georg Simon Ohm, Nurnberg, Germany
[3] Univ Antioquia UdeA, Fac Engn, Calle 70 52-21, Medellin, Colombia
[4] Ludwig Maximilians Univ Munchen, Dept Otorhinolaryngol Head & Neck Surg, Munich, Germany
来源
INTERSPEECH 2021 | 2021年
基金
欧盟地平线“2020”;
关键词
ComParE; Covid-19; recognition; phonetic speech analysis;
D O I
10.21437/Interspeech.2021-1488
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Against the background of the ongoing pandemic, this year's Computational Paralinguistics Challenge featured a classification problem to detect Covid-19 from speech recordings. The presented approach is based on a phonetic analysis of speech samples, thus it enabled us not only to discriminate between Covid and non-Covid samples, but also to better understand how the condition influenced an individual's speech signal. Our deep acoustic model was trained with datasets collected exclusively from healthy speakers. It served as a tool for segmentation and feature extraction on the samples from the challenge dataset. Distinct patterns were found in the embeddings of phonetic classes that have their place of articulation deep inside the vocal tract. We observed profound differences in classification results for development and test splits, similar to the baseline method. We concluded that, based on our phonetic findings, it was safe to assume that our classifier was able to reliably detect a pathological condition located in the respiratory tract. However, we found no evidence to claim that the system was able to discriminate between Covid-19 and other respiratory diseases.
引用
收藏
页码:441 / 445
页数:5
相关论文
共 25 条
[1]   COVID-19: immunopathology and its implications for therapy [J].
Cao, Xuetao .
NATURE REVIEWS IMMUNOLOGY, 2020, 20 (05) :269-270
[2]  
Deshpande G., 2020, ARXIV
[3]  
Garofolo J., 1993, DARPA TIMIT acousticphonetic continuous speech corpus, DOI [10.6028/nist.ir.4930, 10.6028/NIST.IR.4930, DOI 10.6028/NIST.IR.4930]
[4]  
Graves A, 2012, STUD COMPUT INTELL, V385, P1, DOI [10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
[5]  
Graves A., 2006, INT C MACH LEARN
[6]  
Hassan A., 2020, P 2020 IEEE INT C CO
[7]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[8]   Searching for MobileNetV3 [J].
Howard, Andrew ;
Sandler, Mark ;
Chu, Grace ;
Chen, Liang-Chieh ;
Chen, Bo ;
Tan, Mingxing ;
Wang, Weijun ;
Zhu, Yukun ;
Pang, Ruoming ;
Vasudevan, Vijay ;
Le, Quoc V. ;
Adam, Hartwig .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1314-1324
[9]  
Ioffe S, 2015, PR MACH LEARN RES, V37, P448
[10]   Surgical mask detection with deep recurrent phonetic models [J].
Klumpp, Philipp ;
Arias-Vergara, Tomas ;
Camilo Vasquez-Correa, Juan ;
Andrea Perez-Toro, Paula ;
Hoenig, Florian ;
Noeth, Elmar ;
Rafael Orozco-Arroyave, Juan .
INTERSPEECH 2020, 2020, :2057-2061