Biological Cybernetics
Learning internal representations by error propagation
Parallel distributed processing: explorations in the microstructure of cognition, vol. 1
Evolutionary computation: toward a new philosophy of machine intelligence
Evolutionary computation: toward a new philosophy of machine intelligence
Evolution and Optimum Seeking: The Sixth Generation
Evolution and Optimum Seeking: The Sixth Generation
Introduction to Biological and Artificial Neural Networks for Pattern Recognition
Introduction to Biological and Artificial Neural Networks for Pattern Recognition
Stochastic Relaxation, Gibbs Distributions, and the Bayesian Restoration of Images
IEEE Transactions on Pattern Analysis and Machine Intelligence
A clustering algorithm using an evolutionary programming-based approach
Pattern Recognition Letters
P3: a code and design conventions preprocessor for Java
Software—Practice & Experience
A Memetic Pareto Evolutionary Approach to Artificial Neural Networks
AI '01 Proceedings of the 14th Australian Joint Conference on Artificial Intelligence: Advances in Artificial Intelligence
Advances in evolutionary computing
Nonlinear feature extraction using a neuro genetic hybrid
GECCO '05 Proceedings of the 7th annual conference on Genetic and evolutionary computation
Structured genetic algorithm representations for neural network evolution
AIAP'07 Proceedings of the 25th conference on Proceedings of the 25th IASTED International Multi-Conference: artificial intelligence and applications
Comparison of Stochastic Global Optimization Methods to Estimate Neural Network Weights
Neural Processing Letters
Prediction of R5, X4, and R5X4 HIV-1 Coreceptor Usage with Evolved Neural Networks
IEEE/ACM Transactions on Computational Biology and Bioinformatics (TCBB)
CEC'09 Proceedings of the Eleventh conference on Congress on Evolutionary Computation
Optimal design of neural nets using hybrid algorithms
PRICAI'00 Proceedings of the 6th Pacific Rim international conference on Artificial intelligence
Determining regularization parameters for derivative free neural learning
MLDM'05 Proceedings of the 4th international conference on Machine Learning and Data Mining in Pattern Recognition
An evolutionary artificial neural networks approach for breast cancer diagnosis
Artificial Intelligence in Medicine
A hybrid algorithm for artificial neural network training
Engineering Applications of Artificial Intelligence
Hi-index | 0.00 |
This article investigates three potential neural network training algorithms in processing active sonar returns. Our tests show that although all three methods can generate reasonable probabilities of detection and false alarm in discriminating between man-made objects and background events, the stochastic training methods of simulated annealing and evolutionary programming can outperform back propagation.Neural networks are parallel structures made up of nonlinear processing nodes that are connected by fixed or variable weights. They can provide arbitrarily complex measurable decision mappings and are well suited for use in pattern recognition. Neural network paradigms generally fall into two categories: unsupervised and supervised learning.Unsupervised learning networks use the neural network topology as a self-organizing clusterer where decision regions are created with respect to the similarity of the input exemplars to other previously observed exemplars. The network adapts its outputs to minimize a function of the spacing between the elements in each organized cluster of data points. This paradigm requires no a priori clustering information.Supervised learning networks operate on input exemplars that are associated with a desired output. Such learning involves iteratively training the neural network to approximate this mapping. System designers often use multilayer perceptrons (see Figure 1) in supervised learning applications because they can perform arbitrarily complex decision mapping. While this article restricts itself to supervised learning networks, the stochastic training methods investigated also apply to other topologies and unsupervised algorithms.Multilayer perceptrons allow for continuous-valued inputs and outputs. The internal layers transform the input vectors into the output space. Internal weights and bias terms define the output of the network given the presented exemplar. This mapping of input exemplars on the network topology defines a response surface over an n-dimensional hyperspace where there are n weight and bias terms to be adapted. Various algorithms can be used to search for the set of weights and biases that minimize selected functions of the error between the actual and desired outputs (such as the mean squared error).Typical response surfaces often possess local minima. Optimization techniques based on gradient descent may stagnate at these potentially suboptimal solutions, rendering the network incapable of sufficient performance. Second-order Newton and quasi-Newton methods may also fall prey to such entrapment.Many tricks have been invented for avoiding this problem, such as restarting with a new random set of weights, training with noisy exemplars, and perturbing the weights when they appear to prematurely converge. While these methods may lead to improved solutions, there is no guarantee that such minima will not also be only locally optimal. Further, the same suboptimal solution may be rediscovered, leading to fruitless oscillatory training behavior.Stochastic techniques offer an alternative to conventional gradient methods. Both simulated annealingand simulated evolution can serve to generate weight and bias sets. Neither suffers from entrapment in local minima (that is, they have asymptotic global, as opposed to only local, convergence properties).This article assesses three methods of training neural networks, (1) back propagation, (2) simulated annealing, and (3) evolutionary programming, in terms of their performance in distinguishing between sonar returns that reverberate from a man-made metal sphere and those caused by sea mounts, fish and plant life, background noise, or similar anomalies. Although we focus on sonar signal processing, the results extend to more general applications of pattern classification.