Mathematical properties of neuronal TD-rules and differential Hebbian learning: a comparison

  • Authors:
  • Christoph Kolodziejski;Bernd Porr;Florentin Wörgötter

  • Affiliations:
  • University of Göttingen, Bernstein Center for Computational Neuroscience, Bunsenstr. 10, 37073, Göttingen, Germany;University of Glasgow, Department of Electronics and Electrical Engineering, Bunsenstr. 10, GT12 8LT, Glasgow, Scotland;University of Göttingen, Bernstein Center for Computational Neuroscience, Bunsenstr. 10, 37073, Göttingen, Germany

  • Venue:
  • Biological Cybernetics
  • Year:
  • 2008

Quantified Score

Hi-index 0.00

Visualization

Abstract

A confusingly wide variety of temporally asymmetric learning rules exists related to reinforcement learning and/or to spike-timing dependent plasticity, many of which look exceedingly similar, while displaying strongly different behavior. These rules often find their use in control tasks, for example in robotics and for this rigorous convergence and numerical stability is required. The goal of this article is to review these rules and compare them to provide a better overview over their different properties. Two main classes will be discussed: temporal difference (TD) rules and correlation based (differential hebbian) rules and some transition cases. In general we will focus on neuronal implementations with changeable synaptic weights and a time-continuous representation of activity. In a machine learning (non-neuronal) context, for TD-learning a solid mathematical theory has existed since several years. This can partly be transfered to a neuronal framework, too. On the other hand, only now a more complete theory has also emerged for differential Hebb rules. In general rules differ by their convergence conditions and their numerical stability, which can lead to very undesirable behavior, when wanting to apply them. For TD, convergence can be enforced with a certain output condition assuring that the δ-error drops on average to zero (output control). Correlation based rules, on the other hand, converge when one input drops to zero (input control). Temporally asymmetric learning rules treat situations where incoming stimuli follow each other in time. Thus, it is necessary to remember the first stimulus to be able to relate it to the later occurring second one. To this end different types of so-called eligibility traces are being used by these two different types of rules. This aspect leads again to different properties of TD and differential Hebbian learning as discussed here. Thus, this paper, while also presenting several novel mathematical results, is mainly meant to provide a road map through the different neuronally emulated temporal asymmetrical learning rules and their behavior to provide some guidance for possible applications.