Accelerated training of max-margin Markov networks with kernels

  • Authors:
  • Xinhua Zhang;Ankan Saha;S. V. N. Vishwanathan

  • Affiliations:
  • Department of Computing Science, University of Alberta, Edmonton, AB T6G2E8, Canada;Department of Computer Science, University of Chicago, Chicago, IL 60637, USA;Department of Statistics and Computer Science, Purdue University, West Lafayette, IN 47907, USA

  • Venue:
  • Theoretical Computer Science
  • Year:
  • 2014

Quantified Score

Hi-index 5.23

Visualization

Abstract

Structured output prediction is an important machine learning problem both in theory and practice, and the max-margin Markov network (M^3N) is an effective approach. All state-of-the-art algorithms for optimizing M^3N objectives take at least O(1/@e) number of iterations to find an @e accurate solution. Nesterov [1] broke this barrier by proposing an excessive gap reduction technique (EGR) which converges in O(1/@e) iterations. However, it is restricted to Euclidean projections which consequently requires an intractable amount of computation for each iteration when applied to solve M^3N. In this paper, we show that by extending EGR to Bregman projection, this faster rate of convergence can be retained, and more importantly, the updates can be performed efficiently by exploiting graphical model factorization. Further, we design a kernelized procedure which allows all computations per iteration to be performed at the same cost as the state-of-the-art approaches.