What size net gives valid generalization?
Neural Computation
Neurocomputing
The gap between abstract and concrete results in machine learning
Journal of Experimental & Theoretical Artificial Intelligence
Neural networks and the bias/variance dilemma
Neural Computation
Original Contribution: Stacked generalization
Neural Networks
Machine learning, neural and statistical classification
Machine learning, neural and statistical classification
Neural Networks for Pattern Recognition
Neural Networks for Pattern Recognition
Learning in design: From characterizing dimensions to working systems
Artificial Intelligence for Engineering Design, Analysis and Manufacturing
Soft computing in engineering design - A review
Advanced Engineering Informatics
Advances in Engineering Software
Hi-index | 0.00 |
In the course of data modelling, many models could be created. Much work has been done on formulating guidelines for model selection. However, by and large, these guidelines are conservative or too specific. Instead of using general guidelines, models could be selected for a particular task based on statistical tests. When selecting one model, others are discarded. Instead of losing potential sources of information, models could be combined to yield better performance. We review the basics of model selection and combination and discuss their differences. Two examples of opportunistic and principled combinations are presented. The first demonstrates that mediocre quality models could be combined to yield significantly better performance. The latter is the main contribution of the paper; it describes and illustrates a novel heuristic approach called the SG(k-NN) ensemble for the generation of good-quality and diverse models that can even improve excellent quality models.