A resource-allocating network for function interpolation
Neural Computation
A function estimation approach to sequential learning with neural networks
Neural Computation
The nature of statistical learning theory
The nature of statistical learning theory
Machine learning, neural and statistical classification
Machine learning, neural and statistical classification
Data clustering using a model granular magnet
Neural Computation
Making large-scale support vector machine learning practical
Advances in kernel methods
The recursive deterministic perceptron neural network
Neural Networks
An Introduction to Neural Networks
An Introduction to Neural Networks
Radial Basis Function Neural Networks with Sequential Learning
Radial Basis Function Neural Networks with Sequential Learning
Statistical Control of RBF-like Networks for Classification
ICANN '97 Proceedings of the 7th International Conference on Artificial Neural Networks
Radial Basis Functions
Expert Systems with Applications: An International Journal
Letters: Convex incremental extreme learning machine
Neurocomputing
Regularization in the selection of radial basis function centers
Neural Computation
Building RBF neural network topology through potential functions
ICANN/ICONIP'03 Proceedings of the 2003 joint international conference on Artificial neural networks and neural information processing
Evolving hierarchical RBF neural networks for breast cancer detection
ICONIP'06 Proceedings of the 13th international conference on Neural information processing - Volume Part III
Kernel orthonormalization in radial basis function neural networks
IEEE Transactions on Neural Networks
IEEE Transactions on Neural Networks
A generalized growing and pruning RBF (GGAP-RBF) neural network for function approximation
IEEE Transactions on Neural Networks
Hi-index | 0.00 |
In this paper, we propose an architecture and learning algorithm for a growing neural network. Drawing inspiration from the idea of electrical potentials, we develop a classifier based on a set of synthesised potential fields over the domain of input space using symmetrical functions (kernels). We propose a multilayer, multiclass potential function generators classifier (POFGEC) utilising growing architecture and a training algorithm to sequentially add potential functions created by the training patterns, if the addition improves the NN classification performance. We also present a pruning algorithm to achieve compact architecture. POFGEC incorporates the electrical potentials concept in the two main neural net building blocks: potential function generators (PFGs) and potential function entities (PFEs), which perform a non-linear transformation of the input data and create the decision rules by constructing the cumulative potential functions and adjusting the weights. The implementation of the presented method with several datasets demonstrates its capabilities in generating classification solutions for datasets of various shapes independent from the number of predefined classes. We also offer substantial comparative analysis with other known approaches in order to fully illustrate the capabilities of the proposed method and its relation with other existing techniques.