Band importance for speech-in-speech recognition

被引:9
作者
Buss, Emily [1 ]
Bosen, Adam [2 ]
机构
[1] Univ N Carolina, Dept Otolaryngol Head & Neck Surg, Chapel Hill, NC 27599 USA
[2] Boys Town Natl Res Hosp, Ctr Hearing Res, Omaha, NE 68131 USA
来源
JASA EXPRESS LETTERS | 2021年 / 1卷 / 08期
基金
美国国家卫生研究院;
关键词
FREQUENCY IMPORTANCE FUNCTIONS; FUNDAMENTAL-FREQUENCY; INTELLIGIBILITY; SENTENCES; NOISE; PERCEPTION; LISTENERS; WORDS; MODEL;
D O I
10.1121/10.0005762
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Predicting masked speech perception typically relies on estimates of the spectral distribution of cues supporting recognition. Current methods for estimating band importance for speech-in-noise use filtered stimuli. These methods are not appropriate for speech-in-speech because filtering can modify stimulus features affecting auditory stream segregation. Here, band importance is estimated by quantifying the relationship between speech recognition accuracy for full-spectrum speech and the target-to-masker ratio by channel at the output of an auditory filterbank. Preliminary results provide support for this approach and indicate that frequencies below 2kHz may contribute more to speech recognition in two-talker speech than in speech-shaped noise.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] Does Sentence-Level Coarticulation Affect Speech Recognition in Noise or a Speech Masker?
    Jett, Brandi
    Buss, Emily
    Best, Virginia
    Oleson, Jacob
    Calandruccio, Lauren
    [J]. JOURNAL OF SPEECH LANGUAGE AND HEARING RESEARCH, 2021, 64 (04): : 1390 - 1403
  • [42] Measuring time-frequency importance functions of speech with bubble noise
    Mandel, Michael I.
    Yoho, Sarah E.
    Healy, Eric W.
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2016, 140 (04) : 2542 - 2553
  • [43] Perception of voice cues and speech-in-speech by children with prelingual single-sided deafness and a cochlear implant
    Arras, Tine
    Rachman, Laura
    van Wieringen, Astrid
    Baskent, Deniz
    [J]. HEARING RESEARCH, 2024, 454
  • [44] The importance of phase in speech enhancement
    Paliwal, Kuldip
    Wojcicki, Kamil
    Shannon, Benjamin
    [J]. SPEECH COMMUNICATION, 2011, 53 (04) : 465 - 494
  • [45] Measuring Speech Recognition With a Matrix Test Using Synthetic Speech
    Nuesse, Theresa
    Wiercinski, Bianca
    Brand, Thomas
    Holube, Inga
    [J]. TRENDS IN HEARING, 2019, 23
  • [46] CyTex: Transforming speech to textured images for speech emotion recognition
    Bakhshi, Ali
    Harimi, Ali
    Chalup, Stephan
    [J]. SPEECH COMMUNICATION, 2022, 139 : 62 - 75
  • [47] The Optimal Speech-to-Background Ratio for Balancing Speech Recognition With Environmental Sound Recognition
    Johnson, Eric M.
    Healy, Eric W.
    [J]. EAR AND HEARING, 2024, 45 (06) : 1444 - 1460
  • [48] The Recognition of Whispered Speech in Real-Time
    Hendrickson, Kristi
    Ernest, Danielle
    [J]. EAR AND HEARING, 2022, 43 (02) : 554 - 562
  • [49] Applications of automatic speech recognition and text-to-speech technologies for hearing assessment: a scoping review
    Fatehifar, Mohsen
    Schlittenlacher, Josef
    Almufarrij, Ibrahim
    Wong, David
    Cootes, Tim
    Munro, Kevin J.
    [J]. INTERNATIONAL JOURNAL OF AUDIOLOGY, 2024,
  • [50] Comparing human and automatic speech recognition in simple and complex acoustic scenes
    Spille, Constantin
    Kollmeier, Birger
    Meyer, Bernd T.
    [J]. COMPUTER SPEECH AND LANGUAGE, 2018, 52 : 123 - 140