Dysarthric speech classification from coded telephone speech using glottal features

被引:32
|
作者
Narendra, N. P. [1 ]
Alku, Paavo [1 ]
机构
[1] Aalto Univ, Dept Signal Proc & Acoust, Espoo 00076, Finland
基金
芬兰科学院;
关键词
Dysarthric speech; Glottal parameters; Glottal source estimation; Glottal inverse filtering; OpenSMILE; Support vector machines; Telemonitoring; PARKINSONS-DISEASE; INTELLIGIBILITY; DATABASE; MODELS; VOICE;
D O I
10.1016/j.specom.2019.04.003
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This paper proposes a new dysarthric speech classification method from coded telephone speech using glottal features. The proposed method utilizes glottal features, which are efficiently estimated from coded telephone speech using a recently proposed deep neural net-based glottal inverse filtering method. Two sets of glottal features were considered: (1) time- and frequency-domain parameters and (2) parameters based on principal component analysis (PCA). In addition, acoustic features are extracted from coded telephone speech using the openSMILE toolkit. The proposed method utilizes both acoustic and glottal features extracted from coded speech utterances and their corresponding dysarthric/healthy labels to train support vector machine classifiers. Separate classifiers are trained using both individual, and the combination of glottal and acoustic features. The coded telephone speech used in the experiments is generated using the adaptive multi-rate codec, which operates in two transmission bandwidths: narrowband (300 Hz - 3.4 kHz) and wideband (50 Hz - 7 kHz). The experiments were conducted using dysarthric and healthy speech utterances of the TORGO and universal access speech (UA-Speech) databases. Classification accuracy results indicated the effectiveness of glottal features in the identification of dysarthria from coded telephone speech. The results also showed that the glottal features in combination with the openSMILE-based acoustic features resulted in improved classification accuracies, which validate the complementary nature of glottal features. The proposed dysarthric speech classification method can potentially be employed in telemonitoring application for identifying the presence of dysarthria from coded telephone speech.
引用
收藏
页码:47 / 55
页数:9
相关论文
共 50 条
  • [21] Estimating the spectral tilt of the glottal source from telephone speech using a deep neural network
    Jokinen, Emma
    Alku, Paavo
    JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2017, 141 (04): : EL327 - EL330
  • [22] Estimation of Glottal Closure Instants from Telephone Speech using a Group Delay-Based Approach that Considers Speech Signal as a Spectrum
    Rachel, G. Anushiya
    Vijayalakshmi, P.
    Nagarajan, T.
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 1181 - 1185
  • [23] Classification of Emotions from Speech using Implicit Features
    Srivastava, Mohit
    Agarwal, Anupam
    2014 9TH INTERNATIONAL CONFERENCE ON INDUSTRIAL AND INFORMATION SYSTEMS (ICIIS), 2014, : 266 - 271
  • [24] Emotion recognition from telephone speech using acoustic and nonlinear features
    Bedoya-Jaramillo, S.
    Orozco-Arroyave, J. R.
    Arias-Londono, J. D.
    Vargas-Bonilla, J. F.
    2013 47TH INTERNATIONAL CARNAHAN CONFERENCE ON SECURITY TECHNOLOGY (ICCST), 2013,
  • [25] Using speech rhythm knowledge to improve dysarthric speech recognition
    Selouani, S. -A.
    Dahmani, H.
    Amami, R.
    Hamam, H.
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2012, 15 (01) : 57 - 64
  • [26] EXPLORING ARTICULATORY CHARACTERISTICS OF CANTONESE DYSARTHRIC SPEECH USING DISTINCTIVE FEATURES
    Wong, Ka Ho
    Yeung, Wing Sum
    Yeung, Yu Ting
    Meng, Helen
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 6495 - 6499
  • [27] Automatic Classification of Speech Dysarthric Intelligibility Levels Using Textual Feature
    Alharbi, Ghadeer F.
    Alamri, Najwa K.
    Sabbeh, Sahar F.
    IEEE ACCESS, 2025, 13 : 39982 - 39992
  • [28] Data Augmentation using Healthy Speech for Dysarthric Speech Recognition
    Vachhani, Bhavik
    Bhat, Chitralekha
    Kopparapu, Sunil Kumar
    19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 471 - 475
  • [29] Speech intelligibility of dysarthric speech: human scores and acoustic-phonetic features
    Xue, Wei
    van Hout, Roeland
    Boogmans, Fleur
    Ganzeboom, Mario
    Cucchiarini, Catia
    Strik, Helmer
    INTERSPEECH 2021, 2021, : 2911 - 2915
  • [30] Recognising Emotions in Dysarthric Speech Using Typical Speech Data
    Alhinti, Lubna
    Cunningham, Stuart
    Christensen, Heidi
    INTERSPEECH 2020, 2020, : 4821 - 4825