Email updates

Keep up to date with the latest news and content from BMC Neuroscience and BioMed Central.

This article is part of the supplement: Eighteenth Annual Computational Neuroscience Meeting: CNS*2009

Open Access Poster presentation

Sharpening projections

Matthew Cook1*, Florian Jug2 and Christoph Krautz2

Author Affiliations

1 Institute of Neuroinformatics, University of Zurich and ETH Zurich, Zurich, Switzerland

2 Institute of Theoretical Computer Science, ETH Zurich, Zurich, Switzerland

For all author emails, please log on.

BMC Neuroscience 2009, 10(Suppl 1):P214  doi:10.1186/1471-2202-10-S1-P214


The electronic version of this article is the complete one and can be found online at: http://www.biomedcentral.com/1471-2202/10/S1/P214


Published:13 July 2009

© 2009 Cook et al; licensee BioMed Central Ltd.

Poster presentation

It is known that neuronal arbors can project roughly topographically from their source to their target area [1]. However, developmental rules in the cortex often provide only roughly reciprocal connections through chemical cues governing the physical location of distant axonal arbors, leaving open the question of whether more precise reciprocal connectivity is possible in these cases [2,3]. We address this question of whether it is possible for a biologically plausible learning rule to adjust the synaptic weights so that the projection effectively becomes more precise than the anatomy alone provides. We have discovered a biologically plausible set of learning rules that can adjust the synaptic weights so that precisely reciprocal ones are strengthened while others are weakened, thus effectively increasing the specificity of the projections.

The question introduced above can be generalized to any number of areas connected in a feed-forward cycle with feedback only coming as a result of information traveling around the entire cycle. We examined the cases of two or three areas connected in a directed cycle. Each area was represented by a pool of linear threshold units having a sigmoidal threshold function, with inter-area weights initialized to random weight matrices, corresponding to the worst case of a neuronal arbor completely covering the target area. For the dynamics of the unit activities and the weights, we combined three techniques that plausibly have biological counterparts: winner-take-all circuitry [4,5], activity regulation [6], and Hebbian learning [7,8]. On the shortest time scale, winner-take-all circuitry within each pool ensures that the current configuration of activity consists of a clearly localized region of activation. On an intermediate time scale, activity regulation within each unit dampens the likelihood for units to win the winner-take-all competition if they have been highly active but increases their chance to win if they have been inactive, thereby enforcing fair use of the units. On the slowest time scale, Hebbian learning increases the weights along active cycles, making them more likely to recur. Over time, this encourages specific cycles to be strengthened, while the activity regulation ensures that these specific cycles cover all of the units in each pool.

Depending on the combinations of parameters used, it was possible to generate two kinds of reciprocal connectivity. In the first kind, strong precise reciprocal connections developed, with weaker strengths for connections close to being reciprocal, and much weaker connections elsewhere. In other words, the product of the weight matrices was a blurred identity matrix. In the second kind, each area partitioned itself into distinct subgroups, and then the subgroups formed fully connected cycles. In this case, the product of the matrices was a crisp block identity matrix. Both forms of connectivity can be useful within larger artificial architectures and we claim they could easily occur in brains.

Acknowledgements

The authors, who are listed alphabetically, would like to thank ETH Research Grant ETH-23 08-1 and EU Project Grant FET-IP-216593.

References

  1. Essen D, Zeki S: The topographic organization of rhesus monkey prestriate cortex.

    J Physiol 1978, 277:193-226. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  2. Felleman D, Essen D: Distributed hierarchical processing in the primate cerebral cortex.

    Cereb Cortex 1991, 1:1-47. PubMed Abstract | Publisher Full Text OpenURL

  3. Price D, Kennedy H, Dehay C, Zhou L, Mercier M: The development of cortical connections.

    Eur J Neurosci 2006, 23:910-920. PubMed Abstract | Publisher Full Text OpenURL

  4. Dayan P, Abbott L: Theoretical neuroscience: computational and mathematical modeling of neural systems. London: MIT Press; 2001. OpenURL

  5. Hertz J, Krogh A, Palmer R: Introduction to the theory of neural computation. Boulder: Westview Press; 1991. OpenURL

  6. Turrigiano G, Nelson S: Homeostatic plasticity in the developing nervous system.

    Nat Rev Neurosci 2004, 5:97-107. PubMed Abstract | Publisher Full Text OpenURL

  7. Buonomano D, Merzenich M: Cortical plasticity: From synapses to maps.

    Annu Rev Neurosci 1998, 21:149-186. PubMed Abstract | Publisher Full Text OpenURL

  8. Hebb D: The organization of behavior: A neuropsychological theory. New York: Wiley; 1949. OpenURL