A Comparison of Multi-label Feature Selection Methods using the Problem Transformation Approach

  • Authors:
  • Newton SpolaôR;Everton Alvares Cherman;Maria Carolina Monard;Huei Diana Lee

  • Affiliations:
  • Laboratory of Computational Intelligence, Institute of Mathematics and Computer Science, University of São Paulo, 13560-970 São Carlos, SP, Brazil;Laboratory of Computational Intelligence, Institute of Mathematics and Computer Science, University of São Paulo, 13560-970 São Carlos, SP, Brazil;Laboratory of Computational Intelligence, Institute of Mathematics and Computer Science, University of São Paulo, 13560-970 São Carlos, SP, Brazil;Laboratory of Bioinformatics, Western Paraná State University, 85867-900 Foz do Iguaçu, PR, Brazil

  • Venue:
  • Electronic Notes in Theoretical Computer Science (ENTCS)
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Feature selection is an important task in machine learning, which can effectively reduce the dataset dimensionality by removing irrelevant and/or redundant features. Although a large body of research deals with feature selection in single-label data, in which measures have been proposed to filter out irrelevant features, this is not the case for multi-label data. This work proposes multi-label feature selection methods which use the filter approach. To this end, two standard multi-label feature selection approaches, which transform the multi-label data into single-label data, are used. Besides these two problem transformation approaches, we use ReliefF and Information Gain to measure the goodness of features. This gives rise to four multi-label feature selection methods. A thorough experimental evaluation of these methods was carried out on 10 benchmark datasets. Results show that ReliefF is able to select fewer features without diminishing the quality of the classifiers constructed using the features selected.