Feature selection stability assessment based on the Jensen-Shannon divergence

  • Authors:
  • Roberto Guzmán-Martínez;Rocío Alaiz-Rodríguez

  • Affiliations:
  • Servicio de Informatica y Comunicaciones, Universidad de León, León, Spain;Dpto. de Ingeniería Eléctrica y de Sistemas, Universidad de Leon, León, Spain

  • Venue:
  • ECML PKDD'11 Proceedings of the 2011 European conference on Machine learning and knowledge discovery in databases - Volume Part I
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

Feature selection and ranking techniques play an important role in the analysis of high-dimensional data. In particular, their stability becomes crucial when the feature importance is later studied in order to better understand the underlying process. The fact that a small change in the dataset may affect the outcome of the feature selection/ranking algorithm has been long overlooked in the literature. We propose an information-theoretic approach, using the Jensen-Shannon divergence to assess this stability (or robustness). Unlike other measures, this new metric is suitable for different algorithm outcomes: full ranked lists, partial sublists (top-k lists) as well as the least studied partial ranked lists. This generalized metric attempts to measure the disagreement among a whole set of lists with the same size, following a probabilistic approach and being able to give more importance to the differences that appear at the top of the list. We illustrate and compare it with popular metrics like the Spearman rank correlation and the Kuncheva's index on feature selection/ ranking outcomes artificially generated and on an spectral fat dataset with different filter-based feature selectors.