Visual Speech Benefit in Clear and Degraded Speech Depends on the Auditory Intelligibility of the Talker and the Number of Background Talkers

被引:8
作者
Blackburn, Catherine L. [1 ]
Kitterick, Padraig T. [2 ,3 ]
Jones, Gary [1 ]
Sumner, Christian J. [1 ,4 ]
Stacey, Paula C. [1 ]
机构
[1] Nottingham Trent Univ, Dept Psychol, 50 Shakespeare St, Nottingham NG1 4FQ, England
[2] Nottingham Biomed Res Ctr, Nottingham, England
[3] Univ Nottingham, Sch Med, Div Clin Neurosci, Nottingham, England
[4] Med Res Council Inst Hearing Res, Nottingham, England
关键词
speech in noise; visual speech; speech perception; integration; degraded speech; COCHLEAR IMPLANTATION; PERCEPTION; NOISE; MASKING; RECOGNITION; RECEPTION; SENTENCES; USERS;
D O I
10.1177/2331216519837866
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Perceiving speech in background noise presents a significant challenge to listeners. Intelligibility can be improved by seeing the face of a talker. This is of particular value to hearing impaired people and users of cochlear implants. It is well known that auditory-only speech understanding depends on factors beyond audibility. How these factors impact on the audio-visual integration of speech is poorly understood. We investigated audio-visual integration when either the interfering background speech (Experiment 1) or intelligibility of the target talkers (Experiment 2) was manipulated. Clear speech was also contrasted with sine-wave vocoded speech to mimic the loss of temporal fine structure with a cochlear implant. Experiment 1 showed that for clear speech, the visual speech benefit was unaffected by the number of background talkers. For vocoded speech, a larger benefit was found when there was only one background talker. Experiment 2 showed that visual speech benefit depended upon the audio intelligibility of the talker and increased as intelligibility decreased. Degrading the speech by vocoding resulted in even greater benefit from visual speech information. A single "independent noise" signal detection theory model predicted the overall visual speech benefit in some conditions but could not predict the different levels of benefit across variations in the background or target talkers. This suggests that, similar to audio-only speech intelligibility, the integration of audio-visual speech cues may be functionally dependent on factors other than audibility and task difficulty, and that clinicians and researchers should carefully consider the characteristics of their stimuli when assessing audio-visual integration.
引用
收藏
页数:14
相关论文
共 40 条
[1]  
[Anonymous], 1969, IEEE T ACOUST SPEECH, VAU17, P225
[2]  
[Anonymous], 1991, Detection theory: A user's guide
[3]   The Intelligibility of Interrupted Speech: Cochlear Implant Users and Normal Hearing Listeners [J].
Bhargava, Pranesh ;
Gaudrain, Etienne ;
Baskent, Deniz .
JARO-JOURNAL OF THE ASSOCIATION FOR RESEARCH IN OTOLARYNGOLOGY, 2016, 17 (05) :475-491
[4]   Informational and energetic masking effects in the perception of two simultaneous talkers [J].
Brungart, DS .
JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2001, 109 (03) :1101-1109
[5]   Neural processes underlying perceptual enhancement by visual speech gestures [J].
Callan, DE ;
Jones, JA ;
Munhall, K ;
Callan, AM ;
Kroos, C ;
Vatikiotis-Bateson, E .
NEUROREPORT, 2003, 14 (17) :2213-2218
[6]   Multimodal contribution to speech perception revealed by independent component analysis: a single-sweep EEG case study [J].
Callan, DE ;
Callan, AM ;
Kroos, C ;
Vatikiotis-Bateson, E .
COGNITIVE BRAIN RESEARCH, 2001, 10 (03) :349-353
[7]   Crossmodal processing in the human brain: Insights from functional neuroimaging studies [J].
Calvert, GA .
CEREBRAL CORTEX, 2001, 11 (12) :1110-1123
[8]   An ideal observer analysis of variability in visual-only speech [J].
Conrey, Brianna ;
Gold, Jason M. .
VISION RESEARCH, 2006, 46 (19) :3243-3258
[9]   Experiments on Auditory-Visual Perception of Sentences by Users of Unilateral, Bimodal, and Bilateral Cochlear Implants [J].
Dorman, Michael F. ;
Liss, Julie ;
Wang, Shuai ;
Berisha, Visar ;
Ludwig, Cimarron ;
Natale, Sarah Cook .
JOURNAL OF SPEECH LANGUAGE AND HEARING RESEARCH, 2016, 59 (06) :1505-1519
[10]   Informational masking: Counteracting the effects of stimulus uncertainty by decreasing target-masker similarity [J].
Durlach, NI ;
Mason, CR ;
Shinn-Cunningham, BG ;
Arbogast, TL ;
Colburn, HS ;
Kidd, G .
JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2003, 114 (01) :368-379