Unsupervised neural network models of the ventral visual stream

被引:171
|
作者
Zhuang, Chengxu [1 ]
Yan, Siming [2 ]
Nayebi, Aran [3 ]
Schrimpf, Martin [4 ]
Frank, Michael C. [1 ]
DiCarlo, James J. [4 ]
Yamins, Daniel L. K. [1 ,5 ,6 ]
机构
[1] Stanford Univ, Dept Psychol, Stanford, CA 94305 USA
[2] Univ Texas Austin, Dept Comp Sci, Austin, TX 78712 USA
[3] Stanford Univ, Neurosci PhD Program, Stanford, CA 94305 USA
[4] MIT, Brain & Cognit Sci, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[5] Stanford Univ, Dept Comp Sci, Stanford, CA 94305 USA
[6] Stanford Univ, Wu Tsai Neurosci Inst, Stanford, CA 94305 USA
基金
美国国家科学基金会;
关键词
ventral visual stream; deep neural networks; unsupervised algorithms; RECEPTIVE-FIELDS; AREA V4; RECOGNITION; INFANTS; INFORMATION; SELECTIVITY; FRAMEWORK; RESPONSES; FEATURES; PATHWAY;
D O I
10.1073/pnas.2014196118
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Deep neural networks currently provide the best quantitative models of the response patterns of neurons throughout the primate ventral visual stream. However, such networks have remained implausible as a model of the development of the ventral stream, in part because they are trained with supervised methods requiring many more labels than are accessible to infants during development. Here, we report that recent rapid progress in unsupervised learning has largely closed this gap. We find that neural network models learned with deep unsupervised contrastive embedding methods achieve neural prediction accuracy in multiple ventral visual cortical areas that equals or exceeds that of models derived using today's best supervised methods and that the mapping of these neural network models' hidden layers is neuroanatomically consistent across the ventral stream. Strikingly, we find that these methods produce brainlike representations even when trained solely with real human child developmental data collected from head-mounted cameras, despite the fact that these datasets are noisy and limited. We also find that semisupervised deep contrastive embeddings can leverage small numbers of labeled examples to produce representations with substantially improved error-pattern consistency to human behavior. Taken together, these results illustrate a use of unsupervised learning to provide a quantitative model of a multiarea cortical brain system and present a strong candidate for a biologically plausible computational theory of primate sensory learning.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Unsupervised learning of mid-level visual representations
    Matteucci, Giulio
    Piasini, Eugenio
    Zoccolan, Davide
    CURRENT OPINION IN NEUROBIOLOGY, 2024, 84
  • [42] Neural Network Poisson Models for Behavioural and Neural Spike Train Data
    Khajehnejad, Moein
    Habibollahi, Forough
    Nock, Richard
    Arabzadeh, Ehsan
    Dayan, Peter
    Dezfouli, Amir
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [43] Brain-optimized deep neural network models of human visual areas learn non-hierarchical representations
    St-Yves, Ghislain
    Allen, Emily J.
    Wu, Yihan
    Kay, Kendrick
    Naselaris, Thomas
    NATURE COMMUNICATIONS, 2023, 14 (01)
  • [44] Predicting Identity-Preserving Object Transformations across the Human Ventral Visual Stream
    Mocz, Viola
    Vaziri-Pashkam, Maryam
    Chun, Marvin M.
    Xu, Yaoda
    JOURNAL OF NEUROSCIENCE, 2021, 41 (35): : 7403 - 7419
  • [45] Weak surround suppression of the attentional focus characterizes visual selection in the ventral stream in autism
    Ronconi, Luca
    Gori, Simone
    Federici, Alessandra
    Devita, Maria
    Carna, Serena
    Sali, Maria E.
    Molteni, Massimo
    Casartelli, Luca
    Facoetti, Andrea
    NEUROIMAGE-CLINICAL, 2018, 18 : 912 - 922
  • [46] Representations of Facial Identity Information in the Ventral Visual Stream Investigated with Multivoxel Pattern Analyses
    Goesaert, Elfi
    Op de Beeck, Hans P.
    JOURNAL OF NEUROSCIENCE, 2013, 33 (19): : 8549 - 8558
  • [47] Balanced Increases in Selectivity and Tolerance Produce Constant Sparseness along the Ventral Visual Stream
    Rust, Nicole C.
    DiCarlo, James J.
    JOURNAL OF NEUROSCIENCE, 2012, 32 (30): : 10170 - 10182
  • [48] Memory, perception, and the ventral visual-perirhinal-hippocampal stream: Thinking outside of the boxes
    Bussey, T. J.
    Saksidal, L. M.
    HIPPOCAMPUS, 2007, 17 (09) : 898 - 908
  • [49] Non-accidental properties, metric invariance, and encoding by neurons in a model of ventral stream visual object recognition, VisNet
    Rolls, Edmund T.
    Mills, W. Patrick C.
    NEUROBIOLOGY OF LEARNING AND MEMORY, 2018, 152 : 20 - 31
  • [50] Testing the dorsal stream attention hypothesis: Electrophysiological correlates and the effects of ventral stream damage
    Marrett, Narisa E.
    de-Wit, Lee H.
    Roser, Matthew E.
    Kentridge, Robert W.
    Milner, A. David
    Lambert, Anthony J.
    VISUAL COGNITION, 2011, 19 (09) : 1089 - 1121