Adaptive learning of linguistic hierarchy in a multiple timescale recurrent neural network

  • Authors:
  • Stefan Heinrich;Cornelius Weber;Stefan Wermter

  • Affiliations:
  • Department of Informatics, Knowledge Technology, University of Hamburg, Hamburg, Germany;Department of Informatics, Knowledge Technology, University of Hamburg, Hamburg, Germany;Department of Informatics, Knowledge Technology, University of Hamburg, Hamburg, Germany

  • Venue:
  • ICANN'12 Proceedings of the 22nd international conference on Artificial Neural Networks and Machine Learning - Volume Part I
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

Recent research has revealed that hierarchical linguistic structures can emerge in a recurrent neural network with a sufficient number of delayed context layers. As a representative of this type of network the Multiple Timescale Recurrent Neural Network (MTRNN) has been proposed for recognising and generating known as well as unknown linguistic utterances. However the training of utterances performed in other approaches demands a high training effort. In this paper we propose a robust mechanism for adaptive learning rates and internal states to speed up the training process substantially. In addition we compare the generalisation of the network for the adaptive mechanism as well as the standard fixed learning rates finding at least equal capabilities.