A hierarchy of macrodynamical equations for associative memory
Neural Networks
Notions of associative memory and sparse coding
Neural Networks - 1996 Special issue: four major hypotheses in neuroscience
Extended hopfield network for sequence learning: application to gesture recognition
ICANN'05 Proceedings of the 15th international conference on Artificial Neural Networks: biological Inspirations - Volume Part I
Hi-index | 0.00 |
The synapses of real neural systems seem to have delays. Therefore, it is worthwhile to analyze associative memory models with delayed synapses. Thus, a sequential associative memory model with delayed synapses is discussed, where a discrete synchronous updating rule and a correlation learning rule are employed. Its dynamic properties are analyzed by the statistical neurodynamics. In this paper, we first re-derive the Yanai-Kim theory, which involves macrodynamical equations for the dynamics of the network with serial delay elements. Since their theory needs a computational complexity of O(L4t) to obtain the macroscopic state at time step t where L is the length of delay, it is intractable to discuss the macroscopic properties for a large L limit. Thus, we derive steady state equations using the discrete Fourier transformation, where the computational complexity does not formally depend on L. We show that the storage capacity αC is in proportion to the delay length L with a large L limit, and the proportion constant is 0.195, i.e. αC = 0.195L. These results are supported by computer simulations.