Cross-validation in Fuzzy ARTMAP for large databases
Neural Networks
Application of Cascade Correlation Networks for Structures toChemistry
Applied Intelligence
A Fast Simplified Fuzzy ARTMAP Network
Neural Processing Letters
A New Method to Assist Small Data Set Neural Network Learning
ISDA '06 Proceedings of the Sixth International Conference on Intelligent Systems Design and Applications - Volume 01
High-Throughput Ligand Screening via Preclustering and Evolved Neural Networks
IEEE/ACM Transactions on Computational Biology and Bioinformatics (TCBB)
BIBM '07 Proceedings of the 2007 IEEE International Conference on Bioinformatics and Biomedicine
Soft Computing - A Fusion of Foundations, Methodologies and Applications
Neural-network design for small training sets of high dimension
IEEE Transactions on Neural Networks
IEEE Transactions on Neural Networks
Bio-basis function neural network for prediction of protease cleavage sites in proteins
IEEE Transactions on Neural Networks
Fuzzy ARTMAP with input relevances
IEEE Transactions on Neural Networks
Expert Systems with Applications: An International Journal
Bayesian ARTMAP for regression
Neural Networks
Hi-index | 0.00 |
Obtaining satisfactory results with neural networks depends on the availability of large data samples. The use of small training sets generally reduces performance. Most classical Quantitative Structure-Activity Relationship (QSAR) studies for a specific enzyme system have been performed on small data sets. We focus on the neuro-fuzzy prediction of biological activities of HIV-1 protease inhibitory compounds when inferring from small training sets. We propose two computational intelligence prediction techniques which are suitable for small training sets, at the expense of some computational overhead. Both techniques are based on the FAMR model. The FAMR is a Fuzzy ARTMAP (FAM) incremental learning system used for classification and probability estimation. During the learning phase, each sample pair is assigned a relevance factor proportional to the importance of that pair. The two proposed algorithms in this paper are: 1) The GA-FAMR algorithm, which is new, consists of two stages: a) During the first stage, we use a genetic algorithm (GA) to optimize the relevances assigned to the training data. This improves the generalization capability of the FAMR. b) In the second stage, we use the optimized relevances to train the FAMR. 2) The Ordered FAMR is derived from a known algorithm. Instead of optimizing relevances, it optimizes the order of data presentation using the algorithm of Dagher et al. In our experiments, we compare these two algorithms with an algorithm not based on the FAM, the FS-GA-FNN introduced in . We conclude that when inferring from small training sets, both techniques are efficient, in terms of generalization capability and execution time. The computational overhead introduced is compensated by better accuracy. Finally, the proposed techniques are used to predict the biological activities of newly designed potential HIV-1 protease inhibitors.