C4.5: programs for machine learning
C4.5: programs for machine learning
Machine Learning
Learning to Classify Text Using Support Vector Machines: Methods, Theory and Algorithms
Learning to Classify Text Using Support Vector Machines: Methods, Theory and Algorithms
Choosing Multiple Parameters for Support Vector Machines
Machine Learning
Machine Learning
Cancer classification using gene expression data
Information Systems - Special issue: Data management in bioinformatics
An introduction to variable and feature selection
The Journal of Machine Learning Research
Data Mining: Concepts and Techniques
Data Mining: Concepts and Techniques
Data Mining: Practical Machine Learning Tools and Techniques, Second Edition (Morgan Kaufmann Series in Data Management Systems)
A stability index for feature selection
AIAP'07 Proceedings of the 25th conference on Proceedings of the 25th IASTED International Multi-Conference: artificial intelligence and applications
Direct integration of microarrays for selecting informative genes and phenotype classification
Information Sciences: an International Journal
Bioinformatics with soft computing
IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews
Evolutionary Rough Feature Selection in Gene Expression Data
IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews
Expert Systems with Applications: An International Journal
Hi-index | 0.00 |
Microarray experiments generate quantitative expression measurements for thousands of genes simultaneously, which is useful for phenotype classification of many diseases. Our proposed phenotype classifier is an ensemble method with k-topscoring decision rules. Each rule involves a number of genes, a rank comparison relation among them, and a class label. Current classifiers, which are also ensemble methods, consist of k-top-scoring decision rules. Some of these classifiers fix the number of genes in each rule as a triple or a pair. In this paper, we generalize the number of genes involved in each rule. The number of genes in each rule ranges from 2 to N, respectively. Generalizing the number of genes increases the robustness and the reliability of the classifier for the class prediction of an independent sample. Our algorithm saves resources by combining shorter rules in order to build a longer rule. It converges rapidly toward its high-scoring rule list by implementing several heuristics. The parameter k is determined by applying leave-one-out cross validation to the training dataset.