Unsupervised Learning in LSTM Recurrent Neural Networks

  • Authors:
  • Magdalena Klapper-Rybicka;Nicol N. Schraudolph;Jürgen Schmidhuber

  • Affiliations:
  • -;-;-

  • Venue:
  • ICANN '01 Proceedings of the International Conference on Artificial Neural Networks
  • Year:
  • 2001

Quantified Score

Hi-index 0.00

Visualization

Abstract

While much work has been done on unsupervised learning in feedforward neural network architectures, its potential with (theoretically more powerful) recurrent networks and time-varying inputs has rarely been explored. Here we train Long Short-Term Memory (LSTM) recurrent networks to maximize two information-theoretic objectives for unsupervised learning: Binary Information Gain Optimization (BINGO) and Nonparametric Entropy Optimization (NEO). LSTM learns to discriminate different types of temporal sequences and group them according to a variety of features.