A Convergent Incremental Gradient Method with a Constant Step Size

  • Authors:
  • Doron Blatt;Alfred O. Hero;Hillel Gauchman

  • Affiliations:
  • -;-;-

  • Venue:
  • SIAM Journal on Optimization
  • Year:
  • 2007

Quantified Score

Hi-index 0.00

Visualization

Abstract

An incremental aggregated gradient method for minimizing a sum of continuously differentiable functions is presented. The method requires a single gradient evaluation per iteration and uses a constant step size. For the case that the gradient is bounded and Lipschitz continuous, we show that the method visits infinitely often regions in which the gradient is small. Under certain unimodality assumptions, global convergence is established. In the quadratic case, a global linear rate of convergence is shown. The method is applied to distributed optimization problems arising in wireless sensor networks, and numerical experiments compare the new method with other incremental gradient methods.