Online Text Prediction with Recurrent Neural Networks

  • Authors:
  • Juan Antonio Pérez-Ortiz;Jorge Calera-Rubio;Mikel L. Forcada

  • Affiliations:
  • Departament de Llenguatges i Sistemes Informàtics, Universitat d'Alacant E-03071 Alacant, Spain. E-mail: japerez@dlsi.ua.es ;Departament de Llenguatges i Sistemes Informàtics, Universitat d'Alacant E-03071 Alacant, Spain. E-mail: calera@dlsi.ua.es ;Departament de Llenguatges i Sistemes Informàtics, Universitat d'Alacant E-03071 Alacant, Spain. E-mail: mlf@dlsi.ua.es

  • Venue:
  • Neural Processing Letters
  • Year:
  • 2001

Quantified Score

Hi-index 0.00

Visualization

Abstract

Arithmetic coding is one of the most outstanding techniques for lossless data compression. It attains its good performance with the help of a probability model which indicates at each step the probability of occurrence of each possible input symbol given the current context. The better this model, the greater the compression ratio achieved. This work analyses the use of discrete-time recurrent neural networks and their capability for predicting the next symbol in a sequence in order to implement that model. The focus of this study is on online prediction, a task much harder than the classical offline grammatical inference with neural networks. The results obtained show that recurrent neural networks have no problem when the sequences come from the output of a finite-state machine, easily giving high compression ratios. When compressing real texts, however, the dynamics of the sequences seem to be too complex to be learned online correctly by the net.