Convergence of Approximate and Incremental Subgradient Methods for Convex Optimization

  • Authors:
  • Krzysztof C. Kiwiel

  • Affiliations:
  • -

  • Venue:
  • SIAM Journal on Optimization
  • Year:
  • 2003

Quantified Score

Hi-index 0.00

Visualization

Abstract

We present a unified convergence framework for approximate subgradient methods that covers various stepsize rules (including both diminishing and nonvanishing stepsizes), convergence in objective values, and convergence to a neighborhood of the optimal set. We discuss ways of ensuring the boundedness of the iterates and give efficiency estimates. Our results are extended to incremental subgradient methods for minimizing a sum of convex functions, which have recently been shown to be promising for various large-scale problems, including those arising from Lagrangian relaxation.