Optimizing one-shot learning with binary Synapses

被引:16
作者
Romani, Sandro [1 ]
Amit, Daniel J. [3 ,4 ]
Amit, Yali [2 ]
机构
[1] Univ Roma La Sapienza, I-00185 Rome, Italy
[2] Dept Stat & Comp Sci, Chicago, IL 60637 USA
[3] Hebrew Univ Jerusalem, Racah Inst Phys, IL-91904 Jerusalem, Israel
[4] Univ Rome, INFM, I-00185 Rome, Italy
关键词
D O I
10.1162/neco.2008.10-07-618
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A network of excitatory synapses trained with a conservative version of Hebbian learning is used as a model for recognizing the familiarity of thousands of once-seen stimuli from those never seen before. Such networks were initially proposed for modeling memory retrieval (selective delay activity). We show that the same framework allows the incorporation of both familiarity recognition and memory retrieval, and estimate the network's capacity. In the case of binary neurons, we extend the analysis of Amit and Fusi (1994) to obtain capacity limits based on computations of signal-to-noise ratio of the field difference between selective and non-selective neurons of learned signals. We show that with fast learning (potentiation probability approximately 1), the most recently learned patterns can be retrieved in working memory (selective delay activity). A much higher number of once-seen learned patterns elicit a realistic familiarity signal in the presence of an external field. With potentiation probability much less than 1 (slow learning), memory retrieval disappears, whereas familiarity recognition capacity is maintained at a similarly high level. This analysis is corroborated in simulations. For analog neurons, where such analysis is more difficult, we simplify the capacity analysis by studying the excess number of potentiated synapses above the steady-state distribution. In this framework, we derive the optimal constraint between potentiation and depression probabilities that maximizes the capacity.
引用
收藏
页码:1928 / 1950
页数:23
相关论文
共 33 条
[1]   Model of global spontaneous activity and local structured activity during delay periods in the cerebral cortex [J].
Amit, DJ ;
Brunel, N .
CEREBRAL CORTEX, 1997, 7 (03) :237-252
[2]   Multiple-object working memory - A model for behavioral performance [J].
Amit, DJ ;
Bernacchia, A ;
Yakovlev, V .
CEREBRAL CORTEX, 2003, 13 (05) :435-443
[3]   CONSTRAINTS ON LEARNING IN DYNAMIC SYNAPSES [J].
AMIT, DJ ;
FUSI, S .
NETWORK-COMPUTATION IN NEURAL SYSTEMS, 1992, 3 (04) :443-464
[4]   QUANTITATIVE STUDY OF ATTRACTOR NEURAL NETWORKS RETRIEVING AT LOW SPIKE RATES .2. LOW-RATE RETRIEVAL IN SYMMETRICAL NETWORKS [J].
AMIT, DJ ;
TSODYKS, MV .
NETWORK-COMPUTATION IN NEURAL SYSTEMS, 1991, 2 (03) :275-294
[5]   LEARNING IN NEURAL NETWORKS WITH MATERIAL SYNAPSES [J].
AMIT, DJ ;
FUSI, S .
NEURAL COMPUTATION, 1994, 6 (05) :957-982
[6]   Spike-driven synaptic dynamics generating working memory states [J].
Amit, DJ ;
Mongillo, G .
NEURAL COMPUTATION, 2003, 15 (03) :565-596
[7]   Impact of spatiotemporally correlated images on the structure of memory [J].
Bernacchia, Alberto ;
Amit, Daniel J. .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2007, 104 (09) :3544-3549
[8]   Comparison of computational models of familiarity discrimination in the perirhinal cortex [J].
Bogacz, R ;
Brown, MW .
HIPPOCAMPUS, 2003, 13 (04) :494-524
[9]   The restricted influence of sparseness of coding on the capacity of familiarity discrimination networks [J].
Bogacz, R ;
Brown, MW .
NETWORK-COMPUTATION IN NEURAL SYSTEMS, 2002, 13 (04) :457-485
[10]   Model of familiarity discrimination in the perirhinal cortex [J].
Bogacz, R ;
Brown, MW ;
Giraud-Carrier, C .
JOURNAL OF COMPUTATIONAL NEUROSCIENCE, 2001, 10 (01) :5-23