From statistical inference to a differential learning rule for stochastic neural networks

被引:3
作者
Saglietti, Luca [1 ,2 ]
Gerace, Federica [2 ,3 ]
Ingrosso, Alessandro [4 ]
Baldassi, Carlo [2 ,5 ,6 ]
Zecchina, Riccardo [2 ,5 ,7 ]
机构
[1] Microsoft Res New England, Cambridge, MA USA
[2] Italian Inst Genom Med, Turin, Italy
[3] Politecn Torino, DISAT, Turin, Italy
[4] Columbia Univ, Ctr Theoret Neurosci, New York, NY USA
[5] Bocconi Univ, Bocconi Inst Data Sci & Analyt, Milan, Italy
[6] Ist Nazl Fis Nucl, Turin, Italy
[7] Abdus Salaam Int Ctr Theoret Phys, Trieste, Italy
关键词
associative memory; attractor networks; learning; LONG-TERM POTENTIATION; MEAN-FIELD THEORY; DYNAMICS; MEMORY; MODEL; NEURONS; ORIENTATION; MECHANISMS; ALGORITHM;
D O I
10.1098/rsfs.2018.0033
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Stochastic neural networks are a prototypical computational device able to build a probabilistic representation of an ensemble of external stimuli. Building on the relationship between inference and learning, we derive a synaptic plasticity rule that relies only on delayed activity correlations, and that shows a number of remarkable features. Our delayed-correlations matching (DCM) rule satisfies some basic requirements for biological feasibility: finite and noisy afferent signals, Dale's principle and asymmetry of synaptic connections, locality of the weight update computations. Nevertheless, the DCM rule is capable of storing a large, extensive number of patterns as attractors in a stochastic recurrent neural network, under general scenarios without requiring any modification: it can deal with correlated patterns, a broad range of architectures (with or without hidden neuronal states), one-shot learning with the palimpsest property, all the while avoiding the proliferation of spurious attractors. When hidden units are present, our learning rule can be employed to construct Boltzmann machine-like generative models, exploiting the addition of hidden neurons in feature extraction and classification tasks.
引用
收藏
页数:10
相关论文
共 61 条
[11]  
[Anonymous], COMPUTATIONAL TRADEO
[12]  
[Anonymous], 2015, Advances in neural information processing systems
[13]  
[Anonymous], DALES PRINCIPLE IS N
[14]  
[Anonymous], ECHO STATE NETWORKS
[15]  
[Anonymous], 2002, SPIKING NEURON MODEL
[16]  
[Anonymous], USING FIRING RATE DY
[17]   Inverse Ising Inference Using All the Data [J].
Aurell, Erik ;
Ekeberg, Magnus .
PHYSICAL REVIEW LETTERS, 2012, 108 (09)
[18]  
Azouz R, 1999, J NEUROSCI, V19, P2209
[19]   From inverse problems to learning: a Statistical Mechanics approach [J].
Baldassi, Carlo ;
Gerace, Federica ;
Saglietti, Luca ;
Zecchina, Riccardo .
COMPUTER SIMULATIONS IN PHYSICS AND BEYOND (CSP2017), 2018, 955
[20]   THEORY OF ORIENTATION TUNING IN VISUAL-CORTEX [J].
BENYISHAI, R ;
BAROR, RL ;
SOMPOLINSKY, H .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 1995, 92 (09) :3844-3848