Surrogate maximization/minimization algorithms and extensions

  • Authors:
  • Zhihua Zhang;James T. Kwok;Dit-Yan Yeung

  • Affiliations:
  • Department of Electrical Engineering and Computer Science, University of California, Berkeley, CA;Department of Computer Science and Engineering, Hong Kong University of Science and Technology, Kowloon, Hong Kong;Department of Computer Science and Engineering, Hong Kong University of Science and Technology, Kowloon, Hong Kong

  • Venue:
  • Machine Learning
  • Year:
  • 2007

Quantified Score

Hi-index 0.01

Visualization

Abstract

Surrogate maximization (or minimization) (SM) algorithms are a family of algorithms that can be regarded as a generalization of expectation-maximization (EM) algorithms. An SM algorithm aims at turning an otherwise intractable maximization problem into a tractable one by iterating two steps. The S-step computes a tractable surrogate function to substitute the original objective function and the M-step seeks to maximize this surrogate function. Convexity plays a central role in the S-step. SM algorithms enjoy the same convergence properties as EM algorithms. There are mainly three approaches to the construction of surrogate functions, namely, by using Jensen's inequality, first-order Taylor approximation, and the low quadratic bound principle. In this paper, we demonstrate the usefulness of SM algorithms by taking logistic regression models, AdaBoost and the log-linear model as examples. More specifically, by using different surrogate function construction methods, we devise several SM algorithms, including the standard SM, generalized SM, gradient SM, and quadratic SM algorithms, and their two variants called the conditional surrogate maximization (CSM) and surrogate conditional maximization (SCM) algorithms.