Integrating mechanisms of visual guidance in naturalistic language production

被引:0
作者
Moreno I. Coco
Frank Keller
机构
[1] School of Informatics,Institute for Language, Cognition and Computation
[2] University of Edinburgh,Faculdade de Psicologia
[3] Universidade de Lisboa,undefined
来源
Cognitive Processing | 2015年 / 16卷
关键词
Eye movements; Language production; Scene understanding; Cross-modal processing; Eye–voice span; Structural guidance;
D O I
暂无
中图分类号
学科分类号
摘要
Situated language production requires the integration of visual attention and linguistic processing. Previous work has not conclusively disentangled the role of perceptual scene information and structural sentence information in guiding visual attention. In this paper, we present an eye-tracking study that demonstrates that three types of guidance, perceptual, conceptual, and structural, interact to control visual attention. In a cued language production experiment, we manipulate perceptual (scene clutter) and conceptual guidance (cue animacy) and measure structural guidance (syntactic complexity of the utterance). Analysis of the time course of language production, before and during speech, reveals that all three forms of guidance affect the complexity of visual responses, quantified in terms of the entropy of attentional landscapes and the turbulence of scan patterns, especially during speech. We find that perceptual and conceptual guidance mediate the distribution of attention in the scene, whereas structural guidance closely relates to scan pattern complexity. Furthermore, the eye–voice span of the cued object and its perceptual competitor are similar; its latency mediated by both perceptual and structural guidance. These results rule out a strict interpretation of structural guidance as the single dominant form of visual guidance in situated language production. Rather, the phase of the task and the associated demands of cross-modal cognitive processing determine the mechanisms that guide attention.
引用
收藏
页码:131 / 150
页数:19
相关论文
共 138 条
  • [1] Allopenna P(1998)Tracking the time course of spoken word recognition: evidence for continuous mapping models J Mem Lang 38 419-439
  • [2] Magnuson J(1999)Incremental interpretation at verbs: restricting the domain of subsequent reference Cognition 73 247-264
  • [3] Tanenhaus M(2011)I see what you are saying: The integration of complex speech and scenes during language comprehension Acta Psychol 137 208-216
  • [4] Altmann G(2007)The effect of additional characters on choice of referring expression: everything counts J Mem Lang 56 521-536
  • [5] Kamide Y(2008)Mixed-effects modeling with crossed random effects for subjects and items J Mem Lang 59 390-412
  • [6] Andersson R(2013)Random-effects structure for confirmatory hypothesis testing: keep it maximal J Mem Lang 68 255-278
  • [7] Ferreira F(2003)Minding the clock J Mem Lang 4 653-685
  • [8] Henderson J(2008)Contribution of animacy to grammatical function assignment and word order during production Lingua 2 172-189
  • [9] Arnold J(2006)Watching the eyes when talking about size: an investigation of message formulation and utterance planning J Mem Lang 54 592-609
  • [10] Griffin Z(2009)Viewing task influences eye-movement control during active scene perception J Vis 9 1-15