Fast wrapper feature subset selection in high-dimensional datasets by means of filter re-ranking

  • Authors:
  • Pablo Bermejo;Luis de la Ossa;José A. Gámez;José M. Puerta

  • Affiliations:
  • Department of Computing Systems, Intelligent Systems and Data Mining Laboratory (I3A), University of Castilla-La Mancha, Albacete 02071, Spain;Department of Computing Systems, Intelligent Systems and Data Mining Laboratory (I3A), University of Castilla-La Mancha, Albacete 02071, Spain;Department of Computing Systems, Intelligent Systems and Data Mining Laboratory (I3A), University of Castilla-La Mancha, Albacete 02071, Spain;Department of Computing Systems, Intelligent Systems and Data Mining Laboratory (I3A), University of Castilla-La Mancha, Albacete 02071, Spain

  • Venue:
  • Knowledge-Based Systems
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

This paper deals with the problem of supervised wrapper-based feature subset selection in datasets with a very large number of attributes. Recently the literature has contained numerous references to the use of hybrid selection algorithms: based on a filter ranking, they perform an incremental wrapper selection over that ranking. Though working fine, these methods still have their problems: (1) depending on the complexity of the wrapper search method, the number of wrapper evaluations can still be too large; and (2) they rely on a univariate ranking that does not take into account interaction between the variables already included in the selected subset and the remaining ones. Here we propose a new approach whose main goal is to drastically reduce the number of wrapper evaluations while maintaining good performance (e.g. accuracy and size of the obtained subset). To do this we propose an algorithm that iteratively alternates between filter ranking construction and wrapper feature subset selection (FSS). Thus, the FSS only uses the first block of ranked attributes and the ranking method uses the current selected subset in order to build a new ranking where this knowledge is considered. The algorithm terminates when no new attribute is selected in the last call to the FSS algorithm. The main advantage of this approach is that only a few blocks of variables are analyzed, and so the number of wrapper evaluations decreases drastically. The proposed method is tested over eleven high-dimensional datasets (2400-46,000 variables) using different classifiers. The results show an impressive reduction in the number of wrapper evaluations without degrading the quality of the obtained subset.