BoosTexter: A Boosting-based Systemfor Text Categorization
Machine Learning - Special issue on information retrieval
Correlation-based Feature Selection for Discrete and Numeric Class Machine Learning
ICML '00 Proceedings of the Seventeenth International Conference on Machine Learning
Knowledge Discovery in Multi-label Phenotype Data
PKDD '01 Proceedings of the 5th European Conference on Principles of Data Mining and Knowledge Discovery
Chi2: Feature Selection and Discretization of Numeric Attributes
TAI '95 Proceedings of the Seventh International Conference on Tools with Artificial Intelligence
ML-KNN: A lazy learning approach to multi-label learning
Pattern Recognition
Document Transformation for Multi-label Feature Selection in Text Categorization
ICDM '07 Proceedings of the 2007 Seventh IEEE International Conference on Data Mining
Feature selection for multi-label naive Bayes classification
Information Sciences: an International Journal
Feature selection for multi-label classification problems
IWANN'11 Proceedings of the 11th international conference on Artificial neural networks conference on Advances in computational intelligence - Volume Part I
Graphical feature selection for multilabel classification tasks
IDA'11 Proceedings of the 10th international conference on Advances in intelligent data analysis X
Protein classification with multiple algorithms
PCI'05 Proceedings of the 10th Panhellenic conference on Advances in Informatics
Using mutual information for selecting features in supervised neural net learning
IEEE Transactions on Neural Networks
Hi-index | 0.01 |
This paper introduces a new methodology to perform feature selection in multi-label classification problems. Unlike previous works based on the @g^2 statistics, the proposed approach uses the multivariate mutual information criterion combined with a problem transformation and a pruning strategy. This allows us to consider the possible dependencies between the class labels and between the features during the feature selection process. A way to automatically set the pruning parameter is also proposed, based on the permutation test combined with a resampling strategy. Experiments carried out on both artificial and real-world datasets show the interest of our approach over existing methods.