Speeding Up Feature Subset Selection Through Mutual Information Relevance Filtering

  • Authors:
  • Gert Dijck;Marc M. Hulle

  • Affiliations:
  • Katholieke Universiteit Leuven, Computational Neuroscience Research Group, bus 1021, B-3000 Leuven, Belgium;Katholieke Universiteit Leuven, Computational Neuroscience Research Group, bus 1021, B-3000 Leuven, Belgium

  • Venue:
  • PKDD 2007 Proceedings of the 11th European conference on Principles and Practice of Knowledge Discovery in Databases
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

A relevance filter is proposed which removes features based on the mutual information between class labels and features. It is proven that both feature independence and class conditional feature independence are required for the filter to be statistically optimal. This could be shown by establishing a relationship with the conditional relative entropy framework for feature selection. Removing features at various significance levels as a preprocessing step to sequential forward search leads to a huge increase in speed, without a decrease in classification accuracy. These results are shown based on experiments with 5 high-dimensional publicly available gene expression data sets.