A mathematical theory of energy efficient neural computation and communication

  • Authors:
  • Toby Berger;William B. Levy

  • Affiliations:
  • Department of Electrical and Computer Engineering, University of Virginia, Charlottesville, VA;Department of Neurosurgery and Department of Psychology, University of Virginia, Charlottesville, VA

  • Venue:
  • IEEE Transactions on Information Theory - Special issue on information theory in molecular biology and neuroscience
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

A neuroscience-based mathematical model of how a neuron stochastically processes data and communicates information is introduced and analyzed. Call the neuron in question "neuron j", or just "j". The information j transmits approximately describes the time-varying intensity of the excitation j is continuously experiencing from neural spike trains delivered to its synapses by thousands of other neurons. Neuron j "encodes" this excitation history into a sequence of time instants at which it generates neural spikes of its own. By propagating these spikes along its axon, j acts as a multiaccess, partially degraded broadcast channel with thousands of input and output terminals that employs a time-continuous version of pulse position modulation. The mathematical model features three parameters, m, κ, and b, which largely characterize j as an engine of computation and communication. Each set of values of these parameters corresponds to a long term maximization of the bits j conveys to its targets per joule it expends doing so, which is achieved by distributing the random duration between successive spikes j generates according to a gamma pdf with parameters κ and b and distributing b/A according to a beta probability density with parameters κ and m - κ, where A is the random intensity of the effectively Poisson process of spikes that arrive to the union of all of j's synapses at a randomly chosen time instant.