Linear recursive distributed representations

  • Authors:
  • Thomas Voegtlin;Peter F. Dominey

  • Affiliations:
  • INRIA, Campus Scientifique, B.P. 239, F-54506 Vandoeuvre-Les-Nancy Cedex, France;Institut des Sciences Cognitives, CNRS UMR 5015, 67 boulevard Pinel, 69675 Bron Cedex, France

  • Venue:
  • Neural Networks
  • Year:
  • 2005

Quantified Score

Hi-index 0.00

Visualization

Abstract

Connectionist networks have been criticized for their inability to represent complex structures with systematicity. That is, while they can be trained to represent and manipulate complex objects made of several constituents, they generally fail to generalize to novel combinations of the same constituents. This paper presents a modification of Pollack's Recursive Auto-Associative Memory (RAAM), that addresses this criticism. The network uses linear units and is trained with Oja's rule, in which it generalizes PCA to tree-structured data. Learned representations may be linearly combined, in order to represent new complex structures. This results in unprecedented generalization capabilities. Capacity is orders of magnitude higher than that of a RAAM trained with back-propagation. Moreover, regularities of the training set are preserved in the new formed objects. The formation of new structures displays developmental effects similar to those observed in children when learning to generalize about the argument structure of verbs.