Integration of Partial Information Within and Across Modalities: Contributions to Spoken and Written Sentence Recognition

被引:12
作者
Smith, Kimberly G. [1 ]
Fogerty, Daniel [1 ]
机构
[1] Univ South Carolina, Columbia, SC 29208 USA
来源
JOURNAL OF SPEECH LANGUAGE AND HEARING RESEARCH | 2015年 / 58卷 / 06期
关键词
AUDITORY-VISUAL INTEGRATION; RECEPTION THRESHOLD TEST; WORD-RECOGNITION; INTERRUPTED SPEECH; HEARING-LOSS; INDIVIDUAL-DIFFERENCES; PHONEMIC RESTORATION; PERCEPTUAL SYNTHESIS; COGNITIVE-ABILITIES; NOISE;
D O I
10.1044/2015_JSLHR-H-14-0272
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Purpose: This study evaluated the extent to which partial spoken or written information facilitates sentence recognition under degraded unimodal and multimodal conditions. Method: Twenty young adults with typical hearing completed sentence recognition tasks in unimodal and multimodal conditions across 3 proportions of preservation. In the unimodal condition, performance was examined when only interrupted text or interrupted speech stimuli were available. In the multimodal condition, performance was examined when both interrupted text and interrupted speech stimuli were concurrently presented. Sentence recognition scores were obtained from simultaneous and delayed response conditions. Results: Significantly better performance was obtained for unimodal speech-only compared with text-only conditions across all proportions preserved. The multimodal condition revealed better performance when responses were delayed. During simultaneous responses, participants received equal benefit from speech information when the text was moderately and significantly degraded. The benefit from text in degraded auditory environments occurred only when speech was highly degraded. Conclusions: The speech signal, compared with text, is robust against degradation likely due to its continuous, versus discrete, features. Allowing time for offline linguistic processing is beneficial for the recognition of partial sensory information in unimodal and multimodal conditions. Despite the perceptual differences between the 2 modalities, the results highlight the utility of multimodal speech + text signals.
引用
收藏
页码:1805 / 1817
页数:13
相关论文
共 88 条
[1]   The reverse hierarchy theory of visual perceptual learning [J].
Ahissar, M ;
Hochstein, S .
TRENDS IN COGNITIVE SCIENCES, 2004, 8 (10) :457-464
[2]   Reverse hierarchies and sensory learning [J].
Ahissar, Merav ;
Nahum, Mor ;
Nelken, Israel ;
Hochstein, Shaul .
PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY B-BIOLOGICAL SCIENCES, 2009, 364 (1515) :285-299
[3]  
American National Standards Institute, 2004, SPEC AUD ANSI S3 6 2
[4]  
American National Standards Institute, 1997, METH CALC SPEECH INT
[5]  
[Anonymous], 1969, IEEE T ACOUST SPEECH, VAU17, P225
[6]  
[Anonymous], 2004, HDB MULTISENSORY PRO
[7]  
[Anonymous], 2013, Speech Enhancement: Theory and Practice
[8]  
[Anonymous], 1975, Attention and performance V
[9]   INCREASING THE INTELLIGIBILITY OF SPEECH THROUGH MULTIPLE PHONEMIC RESTORATIONS [J].
BASHFORD, JA ;
RIENER, KR ;
WARREN, RM .
PERCEPTION & PSYCHOPHYSICS, 1992, 51 (03) :211-217
[10]   Use of speech-modulated noise adds strong ''bottom-up'' cues for phonemic restoration [J].
Bashford, JA ;
Warren, RM ;
Brown, CA .
PERCEPTION & PSYCHOPHYSICS, 1996, 58 (03) :342-350