Email updates

Keep up to date with the latest news and content from BMC Neuroscience and BioMed Central.

This article is part of the supplement: Twentieth Annual Computational Neuroscience Meeting: CNS*2011

Open Access Poster presentation

Unifying low-level mechanistic and high-level Bayesian explanations of bistable perceptions: neuronal adaptation for cortical inference

David P Reichert*, Peggy Series and Amos J Storkey

Author Affiliations

School of Informatics, University of Edinburgh, Edinburgh, EH8 9AB, UK

For all author emails, please log on.

BMC Neuroscience 2011, 12(Suppl 1):P320  doi:10.1186/1471-2202-12-S1-P320


The electronic version of this article is the complete one and can be found online at: http://www.biomedcentral.com/1471-2202/12/S1/P320


Published:18 July 2011

© 2011 Reichert et al; licensee BioMed Central Ltd.

This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Poster presentation

Ambiguous images such as the Necker cube evoke bistable perceptions in observers, where the conscious percept alternates between the two possible image interpretations. One classic explanation is that mechanisms like neuronal adaptation underlie the switching phenomenon [1]. On the other hand, one possible high-level explanation [2] is that in performing Bayesian inference, the brain might explore the multimodal posterior distribution over possible image interpretations. For example, sampling from a bimodal distribution could explain the perceptual switching [2], and probabilistic sampling might be a general principle underlying cortical inference [3]. In this computational study of bistable perceptions, we show that both views can be combined: Neuronal adaptation such as changes of neuronal excitability and synaptic depression can be understood to improve the sampling algorithm the brain might perform.

We use Deep Boltzmann Machines (DBMs) as models of cortical processing [4]. DBMs are hierarchal probabilistic neural networks that learn to generate or predict the data they are trained on. For doing inference, one can utilize Markov chain Monte Carlo methods such as Gibbs-sampling, corresponding to the model's neurons switching on stochastically. The model then performs a random walk in state space, exploring the various learned interpretations of an image, thus potentially explaining bistable perceptions (cf. [5]). However, in machine learning one often finds that exploring multi-modal posterior distributions in high-dimensional spaces can be problematic, as models can get stuck in individual modes ('the Markov chain does not mix'). Very recent machine learning work [6,7] has devised a class of methods that alleviate this issue by dynamically changing the model parameters, the connection strengths, during sampling. Interestingly, Welling [6] suggested a potential connection to dynamic synapses in biology.

Here, we make this connection explicit. Using a DBM model that has learned to represent toy images of unambiguous cubes, we show how a sampling algorithm similar to [7] can be understood as modeling dynamic changes to neuronal excitability and synaptic strength, making it possible to switch more easily between modes of the posterior distribution, i.e. the two likely interpretations of the ambiguous Necker cube. Unlike [2], who design an ad-hoc abstract inference process, our approach is based on a concrete hierarchical neural network that has learned to represent the images, and utilizes canonical inference methods, with the additional twist of relating the latter to neuronal adaptation. We also make different hypotheses than [2] w.r.t. where in the brain the perceptual switch is realized (namely, gradually throughout the visual hierarchy) and how probability distributions are represented (one sample at a time). Our study naturally follows up on our earlier work [4], where we showed how similar, homeostatic mechanisms on a slower timescale can cause hallucinations. As a final contribution, we demonstrate how spatial attention directed to specific features of the Necker cube can influence the perceptual switching [8].

References

  1. Blake R: A neural theory of binocular rivalry.

    Psychol Rev 1989, 96:145-167. PubMed Abstract | Publisher Full Text OpenURL

  2. Sundareswara R, Schrater PR: Perceptual multistability predicted by search model for Bayesian decisions.

    J Vis 2008, 8:1-19. PubMed Abstract | Publisher Full Text OpenURL

  3. Fiser J, Berkes B, Orban G, Lengyel M: Statistically optimal perception and learning: from behavior to neural representations.

    Trends Cogn Sci 2010, 14:119-130. PubMed Abstract | Publisher Full Text | PubMed Central Full Text OpenURL

  4. Reichert DP, Series P, Storkey AJ: Hallucinations in Charles Bonnet Syndrome Induced by Homeostasis: a Deep Boltzmann Machine Model.

    Adv Neural Inf Process Syst 2010. PubMed Abstract | PubMed Central Full Text OpenURL

  5. Gershman S, Vul E, Tenenbaum J: Perceptual multistability as Markov chain Monte Carlo inference.

    Adv Neural Inf Process Syst 2009. PubMed Abstract | PubMed Central Full Text OpenURL

  6. Welling M: Herding dynamical weights to learn. In Proceedings of the 26th Annual International Conference on Machine Learning. Montreal, Quebec, Canada: ACM; 2009:1121-1128. OpenURL

  7. Breuleux O, Bengio Y, Vincent P: Unlearning for Better Mixing. Universite de Montreal/DIRO; 2010.

  8. Kawabata N: Attention and depth perception.

    Perception 1986, 15:563-572. PubMed Abstract | Publisher Full Text OpenURL