Some Behavioral and Neurobiological Constraints on Theories of Audiovisual Speech Integration: A Review and Suggestions for New Directions

被引:18
作者
Altieri, Nicholas [1 ]
Pisoni, David B. [2 ]
Townsend, James T. [2 ]
机构
[1] Univ Oklahoma, Dept Psychol, Norman, OK 73072 USA
[2] Indiana Univ, Dept Psychol & Brain Sci, Bloomington, IN 47405 USA
来源
SEEING AND PERCEIVING | 2011年 / 24卷 / 06期
关键词
Audio-visual speech perception; McGurk effect; multisensory enhancement; AUDITORY-VISUAL INTEGRATION; SUPERIOR TEMPORAL SULCUS; CROSSMODAL INTERACTION; NEURONAL OSCILLATIONS; CORTEX ACTIVATION; TALKING FACES; HEARING LIPS; MOTOR THEORY; TIME WINDOW; PERCEPTION;
D O I
10.1163/187847611X595864
中图分类号
Q6 [生物物理学];
学科分类号
071011 ;
摘要
Summerfield (1987) proposed several accounts of audiovisual speech perception, a field of research that has burgeoned in recent years. The proposed accounts included the integration of discrete phonetic features, vectors describing the values of independent acoustical and optical parameters, the filter function of the vocal tract, and articulatory dynamics of the vocal tract. The latter two accounts assume that the representations of audiovisual speech perception are based on abstract gestures, while the former two assume that the representations consist of symbolic or featural information obtained from visual and auditory modalities. Recent converging evidence from several different disciplines reveals that the general framework of Summerfield's feature-based theories should be expanded. An updated framework building upon the feature-based theories is presented. We propose a processing model arguing that auditory and visual brain circuits provide facilitatory information when the inputs are correctly timed, and that auditory and visual speech representations do not necessarily undergo translation into a common code during information processing. Future research on multisensory processing in speech perception should investigate the connections between auditory and visual brain regions, and utilize dynamic modeling tools to further understand the timing and information processing mechanisms involved in audiovisual speech integration. (C) Koninklijke Brill NV, Leiden, 2011
引用
收藏
页码:513 / 539
页数:27
相关论文
共 89 条
  • [1] Not Just for Bimodal Neurons Anymore: The Contribution of Unimodal Neurons to Cortical Multisensory Processing
    Allman, Brian L.
    Keniston, Leslie P.
    Meredith, M. Alex
    [J]. BRAIN TOPOGRAPHY, 2009, 21 (3-4) : 157 - 167
  • [2] Altieri N, 2010, UNIFIED THEORY AUDIO
  • [3] [Anonymous], 1986, RESPONSE TIMES
  • [4] Enhanced visual speech perception in individuals with early-onset hearing impairment
    Auer, Edward T., Jr.
    Bernstein, Lynne E.
    [J]. JOURNAL OF SPEECH LANGUAGE AND HEARING RESEARCH, 2007, 50 (05): : 1157 - 1165
  • [5] Bergeson T.R., 2004, HDB MULTISENSORY PRO, P153
  • [6] Bergeson TR, 2003, VOLTA REV, V103, P347
  • [7] Visual speech perception without primary auditory cortex activation
    Bernstein, LE
    Auer, ET
    Moore, JK
    Ponton, CW
    Don, M
    Singh, M
    [J]. NEUROREPORT, 2002, 13 (03) : 311 - 315
  • [8] Bernstein LE., 2004, HDB MULTISENSORY PRO, P203
  • [9] Bernstein LE, 2005, BLACKW HBK LINGUIST, P79, DOI 10.1002/9780470757024.ch4
  • [10] Quantified acoustic-optical speech signal incongruity identifies cortical sites of audiovisual speech processing
    Bernstein, Lynne E.
    Lu, Zhong-Lin
    Jiang, Jintao
    [J]. BRAIN RESEARCH, 2008, 1242 : 172 - 184