Feature Selection by Approximating the Markov Blanket in a Kernel-Induced Space

  • Authors:
  • Qiang Lou;Zoran Obradovic

  • Affiliations:
  • Center for Information Science and Technology, Temple University, Philadelphia, PA 19122, USA. Email: qianglou@temple.edu, zoran@ist.temple.edu;Center for Information Science and Technology, Temple University, Philadelphia, PA 19122, USA. Email: qianglou@temple.edu, zoran@ist.temple.edu

  • Venue:
  • Proceedings of the 2010 conference on ECAI 2010: 19th European Conference on Artificial Intelligence
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

The proposed feature selection method aims to find a minimum subset of the most informative variables for classification/regression by efficiently approximating the Markov Blanket which is a set of variables that can shield a certain variable from the target. Instead of relying on the conditional independence test or network structure learning, the new method uses Hilbert-Schmidt Independence criterion as a measure of dependence among variables in a kernel-induced space. This allows effective approximation of the Markov Blanket that consists of multiple dependent features rather than being limited to a single feature. In addition, the new method can remove both irrelevant and redundant features at the same time. This method for discovering the Markov Blanket is applicable to both discrete and continuous variables, whereas previous methods cannot be used directly for continuous features and therefore are not applicable to regression problems. Experimental evaluations on synthetic and benchmark classification and regression datasets provide evidence that the new feature selection method can remove useless variables in low and in high dimensional problems more accurately than existing Markov Blanket based alternatives.