Multidimensional visual statistical learning

被引:113
|
作者
Turk-Browne, Nicholas B. [1 ]
Isola, Phillip J. [1 ]
Scholl, Brian J. [1 ]
Treat, Teresa A. [1 ]
机构
[1] Yale Univ, Dept Psychol, New Haven, CT 06520 USA
关键词
statistical learning; feature binding; objects; feature dimensions;
D O I
10.1037/0278-7393.34.2.399
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Recent studies of visual statistical learning (VSL) have demonstrated that statistical regularities in sequences of visual stimuli can be automatically extracted, even without intent or awareness. Despite much work on this topic, however, several fundamental questions remain about the nature of VSL. In particular, previous experiments have not explored the underlying units over which VSL operates. In a sequence of colored shapes, for example, does VSL operate over each feature dimension independently, or over multidimensional objects in which color and shape are bound together? The studies reported here demonstrate that VSL can be both object-based and feature-based, in systematic ways based on how different feature dimensions covary. For example, when each shape covaried perfectly with a particular color, VSL was object-based: Observers expressed robust VSL for colored-shape sub-sequences at test but failed when the test items consisted of monochromatic shapes or color patches. When shape and color pairs were partially decoupled during learning, however, VSL operated over features: Observers expressed robust VSL when the feature dimensions were tested separately. These results suggest that VSL is object-based, but that sensitivity to feature correlations in multidimensional sequences (possibly another form of VSL) may in turn help define what counts as an object.
引用
收藏
页码:399 / 407
页数:9
相关论文
共 50 条
  • [41] Statistical learning of spatiotemporal regularities dynamically guides visual attention across space
    Xu, Zhenzhen
    Theeuwes, Jan
    Los, Sander A.
    ATTENTION PERCEPTION & PSYCHOPHYSICS, 2023, 85 (04) : 1054 - 1072
  • [42] Statistical learning of spatiotemporal regularities dynamically guides visual attention across space
    Zhenzhen Xu
    Jan Theeuwes
    Sander A. Los
    Attention, Perception, & Psychophysics, 2023, 85 : 1054 - 1072
  • [43] Visual statistical learning can drive object-based attentional selection
    Zhao, Libo
    Cosman, Joshua D.
    Vatterott, Daniel B.
    Gupta, Prahlad
    Vecera, Shaun P.
    ATTENTION PERCEPTION & PSYCHOPHYSICS, 2014, 76 (08) : 2240 - 2248
  • [44] Statistical Learning Speeds Visual Search: More Efficient Selection, or Faster Response?
    Wang, Sisi
    Cong, Stanislas Huynh
    Woodman, Geoffrey F.
    JOURNAL OF EXPERIMENTAL PSYCHOLOGY-GENERAL, 2023, 152 (06) : 1723 - 1734
  • [45] Visual learning of statistical relations among nonadjacent features: Evidence for structural encoding
    Barenholtz, Elan
    Tarr, Michael J.
    VISUAL COGNITION, 2011, 19 (04) : 469 - 482
  • [46] Translating visual information into action predictions: Statistical learning in action and nonaction contexts
    Claire D. Monroy
    Sarah A. Gerson
    Sabine Hunnius
    Memory & Cognition, 2018, 46 : 600 - 613
  • [47] Distinct Contributions of Attention and Working Memory to Visual Statistical Learning and Ensemble Processing
    Hall, Michelle G.
    Mattingley, Jason B.
    Dux, Paul E.
    JOURNAL OF EXPERIMENTAL PSYCHOLOGY-HUMAN PERCEPTION AND PERFORMANCE, 2015, 41 (04) : 1112 - 1123
  • [48] Statistical learning in visual search reflects distractor rarity, not only attentional suppression
    Dirk Kerzel
    Chiara Balbiani
    Sarah Rosa
    Stanislas Huynh Cong
    Psychonomic Bulletin & Review, 2022, 29 : 1890 - 1897
  • [49] Recent cross-modal statistical learning influences visual perceptual selection
    Piazza, Elise A.
    Denison, Rachel N.
    Silver, Michael A.
    JOURNAL OF VISION, 2018, 18 (03): : 1 - 12
  • [50] Visual statistical learning can drive object-based attentional selection
    Libo Zhao
    Joshua D. Cosman
    Daniel B. Vatterott
    Prahlad Gupta
    Shaun P. Vecera
    Attention, Perception, & Psychophysics, 2014, 76 : 2240 - 2248