Email updates

Keep up to date with the latest news and content from BMC Neuroscience and BioMed Central.

This article is part of the supplement: Eighteenth Annual Computational Neuroscience Meeting: CNS*2009

Open Access Poster presentation

Self-organization of asymmetric associative networks

Christian Albers and Klaus Pawelzik*

Author Affiliations

Institute for Theoretical Physics, University Bremen, D-28359 Bremen, Germany

For all author emails, please log on.

BMC Neuroscience 2009, 10(Suppl 1):P340  doi:10.1186/1471-2202-10-S1-P340

The electronic version of this article is the complete one and can be found online at:

Published:13 July 2009

© 2009 Albers and Pawelzik; licensee BioMed Central Ltd.


Associative networks serve as basic paradigms for memory retrieval [1-3] and have been used to describe properties of many neuronal structures [4,5]. In contrast, learning via activity dependent adaptation of real synapses is still not well understood. In particular, the Hopfield Model [2] artificially enforces synapses to represent neuronal correlations according to the Hebbian rule that unrealistically leads to symmetric couplings, catastrophic forgetting when large amounts of patterns are to be stored and explosion of weights when the Hebbian rule is applied iteratively. Various modifications of the simple Hebbian rule have been proposed, as for instance the use of global information to control Hebbian contributions to the weight matrix [6] that prohibits catastrophic forgetting and weight explosion. In contrast, we present a local learning rule that in large networks of N neurons can stabilize more than 1.6 N binary non-sparse random patterns (see Figure 1). When applied in an on-line fashion, it leads to retention of the stack of recent patterns without attrition. The synaptic algorithm turns out to be consistent with spike-timing dependent plasticity as observed in hippocampus and cortex. In fact, it resembles the perceptron rule [3] by modifying a synapse only when the postsynaptic neuron changes its activity. We also find that the mutual interaction of network dynamics with weight changes confines synaptic strength to finite values despite the formal instability of the local synaptic learning rule. Taken together our work suggests that stabilization could provide a unifying principle of weight-activity co-evolution, leading to large storage capacity, on-line ability, generalization and extraction of higher order correlations which has testable implications for synaptic dynamics and cortical function.

thumbnailFigure 1. Retrieval of stored patterns depending on the relative number P/N of patterns learned in a network of N neurons.


Many thanks to Misha Tsodyks for encouraging discussions.


  1. Palm G: Neural Assemblies. In An Alternative Approach to Artificial Intelligence. Springer, Berlin, Heidelberg, New York; 1982. OpenURL

  2. Hopfield JJ: Neural networks and physical systems with emergent collective computational abilities.

    Proc NatL Acad Sci USA 1982, 79:2554-2558. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  3. Hertz JA, Krogh A, Palmer RG: Introduction to the Theory of Neural Computation. Addison Wesley; 1991. OpenURL

  4. Myashita Y: Inferior temporal cortex: Where visual perception meets memory.

    Ann Rev Neurosci 1993, 16:245-263. PubMed Abstract | Publisher Full Text OpenURL

  5. Amit DJ, Brunel N, Tsodyks MV: Correlations of cortical Hebbian reverberations: theory versus experiment.

    J Neuroscience 1994, 14:6435-6445. OpenURL

  6. Blumenfeld B, Preminger S, Sagi D, Tsodyks M: Dynamics of memory representations in networks with novelty-facilitated synaptic plasticity.

    Neuron 52:383-394. PubMed Abstract | Publisher Full Text OpenURL