Regularization techniques for learning with matrices

  • Authors:
  • Sham M. Kakade;Shai Shalev-Shwartz;Ambuj Tewari

  • Affiliations:
  • Microsoft Research New England, Cambridge, MA;School of Computer Science and Engineering, The Hebrew University of Jerusalem, Jerusalem, Israel;Department of Computer Science, The University of Texas at Austin, Austin, TX

  • Venue:
  • The Journal of Machine Learning Research
  • Year:
  • 2012

Quantified Score

Hi-index 0.00

Visualization

Abstract

There is growing body of learning problems for which it is natural to organize the parameters into a matrix. As a result, it becomes easy to impose sophisticated prior knowledge by appropriately regularizing the parameters under some matrix norm. This work describes and analyzes a systematic method for constructing such matrix-based regularization techniques. In particular, we focus on how the underlying statistical properties of a given problem can help us decide which regularization function is appropriate. Our methodology is based on a known duality phenomenon: a function is strongly convex with respect to some norm if and only if its conjugate function is strongly smooth with respect to the dual norm. This result has already been found to be a key component in deriving and analyzing several learning algorithms. We demonstrate the potential of this framework by deriving novel generalization and regret bounds for multi-task learning, multi-class learning, and multiple kernel learning.