Incremental training of first order recurrent neural networks to predict a context-sensitive language

  • Authors:
  • Stephan K. Chalup;Alan D. Blair

  • Affiliations:
  • School of Electrical Engineering and Computer Science, The University of Newcastle, Callaghan, NSW 2308, Australia;School of Computer Science and Engineering, The University of New South Wales, Sydney, NSW 2052, Australia

  • Venue:
  • Neural Networks
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

In recent years it has been shown that first order recurrent neural networks trained by gradient-descent can learn not only regular but also simple context-free and context-sensitive languages. However, the success rate was generally low and severe instability issues were encountered. The present study examines the hypothesis that a combination of evolutionary hill climbing with incremental learning and a well-balanced training set enables first order recurrent networks to reliably learn context-free and mildly context-sensitive languages. In particular, we trained the networks to predict symbols in string sequences of the context-sensitive language {anbncn; n ≥ 1}. Comparative experiments with and without incremental learning indicated that incremental learning can accelerate and facilitate training. Furthermore, incrementally trained networks generally resulted in monotonic trajectories in hidden unit activation space, while the trajectories of non- incrementally trained networks were oscillating. The non-incrementally trained networks were more likely to generalise.