An unsupervised learning based LSTM model: a new architecture

  • Authors:
  • Sajjad Mohsin;Fatima Zaka

  • Affiliations:
  • Department of Computer Science, COMSATS Institute of Information Technology, Islamabad, Pakistan;Department of Computer Science, COMSATS Institute of Information Technology, Islamabad, Pakistan

  • Venue:
  • AMERICAN-MATH'11/CEA'11 Proceedings of the 2011 American conference on applied mathematics and the 5th WSEAS international conference on Computer engineering and applications
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Recurrent Neural Networks (RNNs) have shown good results with real-world temporal contextual data, but for input sequences with long time lags they fail. Long Short Term Memory (LSTM) model was built to address the issue of large time lags in input data successfully. However, LSTM found lacking for the tasks pertaining to lower level cognitive processing, specifically, information processing, storage & recall, and also whether they could learn in an unsupervised manner. Sustained Temporal Order Recurrent (STORE) networks are designed to encode the order of temporal data, and then could recall the encoded data in veridical as well as non-veridical order employing unsupervised learning. In this research we have propose the Fusion of supervised learning base LSTM propose by Jurgen and unsupervised learning based STORE proposed by Grossberg. To alternate between two approaches as well as mimicking brain in information processing during sleep time (internal input) we proposed CCS (Consolidation Control Unit), built on an in-depth cognitive foundation, to overcome the inability of LSTM to learn in unsupervised manner and to work with lower level cognitive processing. We conclude by providing experimental proof of the efficiency of proposed model by comparing it with original LSTM model.