Slow stochastic learning with global inhibition: a biological solution to the binary perceptron problem
Introduction
The strength of biological synapses can only vary within a limited range, and there is accumulating evidence that some synapses can only preserve a restricted number of states (some seem to have only two [4]). These constraints have dramatic effects on networks performing as classifiers or as an associative memory. Networks of neurons connected by bounded synapses which cannot be changed by an arbitrarily small amount share the palimpsest property (see e.g. [2]): new patterns overwrite the oldest ones, and only a limited number of patterns can be remembered. The more synapses changed on each stimulus presentation, the faster is forgetting. Moreover, learning to separate two classes of patterns with discrete synaptic weights is a combinatorially hard problem (the ‘binary perceptron problem’, see [1]). Fast forgetting can be avoided by changing only a small fraction of synapses, chosen randomly at each presentation. Stochastic selection permits the classification and memorization of an extensive number of random patterns, even if the number of synaptic states is reduced to two [2]. However, additional mechanisms must be introduced to store more realistic patterns with correlated components. The solution we study here is based on the perceptron learning rule: the synapses are changed with some probability only when the response of the post-synaptic cell is not the desired one. This ‘stop-learning’ property might be the expression of some regulatory synaptic mechanisms or the effect of a reward signal. Together with global inhibition, a small synaptic transition probability and a small neuronal threshold are sufficient to learn and memorize any linearly separable set of patterns.
Section snippets
The model
Neuron model: We consider a single postsynaptic neuron which receives excitatory inputs from N presynaptic neurons, and an inhibitory input which is proportional to the total activity of the N excitatory neurons. The postsynaptic neuron is either active or inactive, depending on whether the total postsynaptic current h is above or below a threshold θ0. The total current is calculated by the weighted sum of the synaptic inputs ξj, , where ξj can take on any value from (and
Results
Given any two sets C± of linearly separable patterns, a neuron endowed with global inhibition and the stochastic learning rule described above will always learn to correctly classify the patterns in a finite number of presentations. The tighter the separation between the two classes C±, the smaller the neuronal threshold θ0, the learning margin δ0, and the learning rate q must be (for simplicity we assume q+=q−=q). More precisely, we assume that there is a separation vector S of length ||S||=N
Conclusions
We have shown that stochastic learning allows a perceptron with binary excitatory weights to converge in a finite number of updates for any separable set of patterns, provided that there is some global inhibition, a small neuronal threshold, and slow learning. These ingredients rescue binary synapses from fast forgetting due to saturation of the potentiation probabilities. They also allow to store as many patterns as in a network with analogue unbounded synapses (proportional to Nα, with α from
Acknowledgements
This work was supported by the EU Grant IST-2001-38099 ALAVLSI and the SNF Grant 3152-065234.01. We thank J. Brader for useful remarks.
References (4)
- et al.
Computers and Intractability: A Guide to the Theory of NP-Completeness
(1999) Hebbian spike-driven synaptic plasticity for learning patterns of mean firing rates
Biol. Cybern.
(2002)
Cited by (5)
Generalization of finite size Boolean perceptrons with genetic algorithms
2008, NeurocomputingCitation Excerpt :Using statistical mechanics techniques, Gardner and Derrida [7] introduced the study of Boolean perceptrons with discrete weights without correlation among input patterns and the corresponding outputs, the so-called random map. For finite size networks learning in binary perceptron has been investigated with different techniques such as the slow stochastic process devised by Senn et al. [19,20] and clipping of continuous-weight perceptrons [5,13,14,17]. Random input–output associations algorithm were investigated by Baldassi et al. [2], who showed that on-line supervised algorithms provides fast learning of random input–output associations, up to close to the theoretical capacity [12].
Are binary synapses superior to graded weight representations in stochastic attractor networks?
2009, Cognitive NeurodynamicsMultiple views of the response of an ensemble of spectro-temporal features support concurrent classification of utterance, prosody, sex and speaker identity
2005, Network: Computation in Neural SystemsConvergence of stochastic learning in perceptrons with binary synapses
2005, Physical Review E - Statistical, Nonlinear, and Soft Matter Physics