Learning sequential structure with the real-time recurrent learning algorithm
International Journal of Neural Systems
The neural network pushdown automation: model, stack and learning simulations
The neural network pushdown automation: model, stack and learning simulations
Neural Computation
Recurrent neural networks can learn to implement symbol-sensitive counting
NIPS '97 Proceedings of the 1997 conference on Advances in neural information processing systems 10
Kalman Filtering and Neural Networks
Kalman Filtering and Neural Networks
Neural Networks: A Comprehensive Foundation
Neural Networks: A Comprehensive Foundation
Learning to Forget: Continual Prediction with LSTM
Neural Computation
LSTM recurrent networks learn simple context-free and context-sensitive languages
IEEE Transactions on Neural Networks
Gradient calculations for dynamic recurrent neural networks: a survey
IEEE Transactions on Neural Networks
Training Recurrent Networks by Evolino
Neural Computation
Kalman filtering for neural prediction of response spectra from mining tremors
Computers and Structures
ICANN '08 Proceedings of the 18th international conference on Artificial Neural Networks, Part I
Error entropy minimization for LSTM training
ICANN'06 Proceedings of the 16th international conference on Artificial Neural Networks - Volume Part I
A multitask approach to continuous five-dimensional affect sensing in natural speech
ACM Transactions on Interactive Intelligent Systems (TiiS) - Special Issue on Affective Interaction in Natural Environments
Hi-index | 0.01 |
The long short-term memory (LSTM) network trained by gradient descent solves difficult problems which traditional recurrent neural networks in general cannot. We have recently observed that the decoupled extended Kalman filter training algorithm allows for even better performance, reducing significantly the number of training steps when compared to the original gradient descent training algorithm. In this paper we present a set of experiments which are unsolvable by classical recurrent networks but which are solved elegantly and robustly and quickly by LSTM combined with Kalman filters.