Boolean Feature Discovery in Empirical Learning
Machine Learning
Experimental comparison of human and machine learning formalisms
Proceedings of the sixth international workshop on Machine learning
Learning hard concepts through constructive induction: framework and rationale
Computational Intelligence
A practical approach to feature selection
ML92 Proceedings of the ninth international workshop on Machine learning
C4.5: programs for machine learning
C4.5: programs for machine learning
Estimating attributes: analysis and extensions of RELIEF
ECML-94 Proceedings of the European conference on machine learning on Machine Learning
Elements of machine learning
A decision-theoretic generalization of on-line learning and an application to boosting
Journal of Computer and System Sciences - Special issue: 26th annual ACM symposium on the theory of computing & STOC'94, May 23–25, 1994, and second annual Europe an conference on computational learning theory (EuroCOLT'95), March 13–15, 1995
Selection of relevant features and examples in machine learning
Artificial Intelligence - Special issue on relevance
A re-examination of text categorization methods
Proceedings of the 22nd annual international ACM SIGIR conference on Research and development in information retrieval
Data mining: practical machine learning tools and techniques with Java implementations
Data mining: practical machine learning tools and techniques with Java implementations
BoosTexter: A Boosting-based Systemfor Text Categorization
Machine Learning - Special issue on information retrieval
Relational Data Mining
Feature Selection Using Rough Sets Theory
ECML '93 Proceedings of the European Conference on Machine Learning
Correlation-based Feature Selection for Discrete and Numeric Class Machine Learning
ICML '00 Proceedings of the Seventeenth International Conference on Machine Learning
Improvements to Platt's SMO Algorithm for SVM Classifier Design
Neural Computation
Spatial associative classification: propositional vs structural approach
Journal of Intelligent Information Systems
Improving multiclass pattern recognition with a co-evolutionary RBFNN
Pattern Recognition Letters
Stable feature selection via dense feature groups
Proceedings of the 14th ACM SIGKDD international conference on Knowledge discovery and data mining
Internet traffic classification demystified: myths, caveats, and the best practices
CoNEXT '08 Proceedings of the 2008 ACM CoNEXT Conference
Consensus group stable feature selection
Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining
Dual-population based coevolutionary algorithm for designing RBFNN with feature selection
Expert Systems with Applications: An International Journal
WSEAS TRANSACTIONS on COMMUNICATIONS
Mining relational association rules for propositional classification
AI*IA'05 Proceedings of the 9th conference on Advances in Artificial Intelligence
Reducing examples in relational learning with bounded-treewidth hypotheses
NFMCP'12 Proceedings of the First international conference on New Frontiers in Mining Complex Patterns
Stable Feature Selection with Minimal Independent Dominating Sets
Proceedings of the International Conference on Bioinformatics, Computational Biology and Biomedical Informatics
Hi-index | 0.00 |
We consider the problem of eliminating redundant Boolean features for a given data set, where a feature is redundant if it separates the classes less well than another feature or set of features. Lavrač et al. proposed the algorithm REDUCE that works by pairwise comparison of features, i.e., it eliminates a feature if it is redundant with respect to another feature. Their algorithm operates in an ILP setting and is restricted to two-class problems. In this paper we improve their method and extend it to multiple classes. Central to our approach is the notion of a neighbourhood of examples: a set of examples of the same class where the number of different features between examples is relatively small. Redundant features are eliminated by applying a revised version of the REDUCE method to each pair of neighbourhoods of different class. We analyse the performance of our method on a range of data sets.