From statistical inference to a differential learning rule for stochastic neural networks

被引:3
作者
Saglietti, Luca [1 ,2 ]
Gerace, Federica [2 ,3 ]
Ingrosso, Alessandro [4 ]
Baldassi, Carlo [2 ,5 ,6 ]
Zecchina, Riccardo [2 ,5 ,7 ]
机构
[1] Microsoft Res New England, Cambridge, MA USA
[2] Italian Inst Genom Med, Turin, Italy
[3] Politecn Torino, DISAT, Turin, Italy
[4] Columbia Univ, Ctr Theoret Neurosci, New York, NY USA
[5] Bocconi Univ, Bocconi Inst Data Sci & Analyt, Milan, Italy
[6] Ist Nazl Fis Nucl, Turin, Italy
[7] Abdus Salaam Int Ctr Theoret Phys, Trieste, Italy
关键词
associative memory; attractor networks; learning; LONG-TERM POTENTIATION; MEAN-FIELD THEORY; DYNAMICS; MEMORY; MODEL; NEURONS; ORIENTATION; MECHANISMS; ALGORITHM;
D O I
10.1098/rsfs.2018.0033
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Stochastic neural networks are a prototypical computational device able to build a probabilistic representation of an ensemble of external stimuli. Building on the relationship between inference and learning, we derive a synaptic plasticity rule that relies only on delayed activity correlations, and that shows a number of remarkable features. Our delayed-correlations matching (DCM) rule satisfies some basic requirements for biological feasibility: finite and noisy afferent signals, Dale's principle and asymmetry of synaptic connections, locality of the weight update computations. Nevertheless, the DCM rule is capable of storing a large, extensive number of patterns as attractors in a stochastic recurrent neural network, under general scenarios without requiring any modification: it can deal with correlated patterns, a broad range of architectures (with or without hidden neuronal states), one-shot learning with the palimpsest property, all the while avoiding the proliferation of spurious attractors. When hidden units are present, our learning rule can be employed to construct Boltzmann machine-like generative models, exploiting the addition of hidden neurons in feature extraction and classification tasks.
引用
收藏
页数:10
相关论文
共 61 条
[1]   Building functional networks of spiking model neurons [J].
Abbott, L. F. ;
DePasquale, Brian ;
Memmesheimer, Raoul-Martin .
NATURE NEUROSCIENCE, 2016, 19 (03) :350-355
[2]  
ACKLEY DH, 1985, COGNITIVE SCI, V9, P147
[3]   A Three-Threshold Learning Rule Approaches the Maximal Capacity of Recurrent Neural Networks [J].
Alemi, Alireza ;
Baldassi, Carlo ;
Brunel, Nicolas ;
Zecchina, Riccardo .
PLOS COMPUTATIONAL BIOLOGY, 2015, 11 (08)
[4]   Model of global spontaneous activity and local structured activity during delay periods in the cerebral cortex [J].
Amit, DJ ;
Brunel, N .
CEREBRAL CORTEX, 1997, 7 (03) :237-252
[5]   INFORMATION-STORAGE IN NEURAL NETWORKS WITH LOW-LEVELS OF ACTIVITY [J].
AMIT, DJ ;
GUTFREUND, H ;
SOMPOLINSKY, H .
PHYSICAL REVIEW A, 1987, 35 (05) :2293-2303
[6]  
Amit DJ., 1989, MODELING BRAIN FUNCT, DOI DOI 10.1017/CBO9780511623257
[7]  
[Anonymous], ADV NEURAL INFORM PR
[8]  
[Anonymous], 2008, 2008002 UTML TR DEP
[9]  
[Anonymous], 2005, AISTATS BRIDGETOWN B
[10]  
[Anonymous], 2012, INTERDISCIP INF SCI, DOI DOI 10.4036/iis.2012.55