Connectionist learning procedures
Artificial Intelligence
Learning internal representations by error propagation
Parallel distributed processing: explorations in the microstructure of cognition, vol. 1
Improving the convergence of the back-propagation algorithm
Neural Networks
Neural Networks for Pattern Recognition
Neural Networks for Pattern Recognition
The effect of learning on life history evolution
Proceedings of the 9th annual conference on Genetic and evolutionary computation
Lifetime learning as a factor in life history evolution
Artificial Life
Evolutionary Fuzzy ARTMAP Neural Networks and their Applications to Fault Detection and Diagnosis
Neural Processing Letters
Hi-index | 0.00 |
Gradient descent training of sigmoidal feed-forward neural networks on binary mappings often gets stuck with someout puts totally wrong. This is because a sum-squared-error cost function leads to weight updates that depend on the derivative of the output sigmoid which goes to zero as the output approaches maximal error. Although it is easy to understand the cause, the best remedy is not so obvious. Common solutions involve modifying the training data, deviating from true gradient descent, or changing the cost function. In general, finding the best learning procedures for particular classes of problem is difficult because each usually depends on a number of interacting parameters that need to be set to optimal values for a fair comparison. In this paper I shall use simulated evolution to optimise all the relevant parameters, and come to a clear conclusion concerning the most efficient approach for learning binary mappings.