Joint covariate selection and joint subspace selection for multiple classification problems

  • Authors:
  • Guillaume Obozinski;Ben Taskar;Michael I. Jordan

  • Affiliations:
  • Department of Statistics, University of California at Berkeley, Berkeley, USA 94720-3860;Department of Computer and Information Science, University of Pennsylvania, Philadelphia, USA 19104-6389;Department of Statistics and Department of Electrical Engineering and Computer Science, University of California at Berkeley, Berkeley, USA 94720-3860

  • Venue:
  • Statistics and Computing
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

We address the problem of recovering a common set of covariates that are relevant simultaneously to several classification problems. By penalizing the sum of 驴 2 norms of the blocks of coefficients associated with each covariate across different classification problems, similar sparsity patterns in all models are encouraged. To take computational advantage of the sparsity of solutions at high regularization levels, we propose a blockwise path-following scheme that approximately traces the regularization path. As the regularization coefficient decreases, the algorithm maintains and updates concurrently a growing set of covariates that are simultaneously active for all problems. We also show how to use random projections to extend this approach to the problem of joint subspace selection, where multiple predictors are found in a common low-dimensional subspace. We present theoretical results showing that this random projection approach converges to the solution yielded by trace-norm regularization. Finally, we present a variety of experimental results exploring joint covariate selection and joint subspace selection, comparing the path-following approach to competing algorithms in terms of prediction accuracy and running time.