Distributed stochastic approximation for constrained and unconstrained optimization

  • Authors:
  • Pascal Bianchi;Jérémie Jakubowicz

  • Affiliations:
  • Telecom ParisTech/CNRS-LTCI, rue Barrault, Paris Cedex - France;Telecom ParisTech/CNRS-LTCI, rue Barrault, Paris Cedex - France

  • Venue:
  • Proceedings of the 5th International ICST Conference on Performance Evaluation Methodologies and Tools
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

In this paper, we analyze the convergence of a distributed Robbins-Monro algorithm for both constrained and unconstrained optimization in multi-agent systems. The algorithm searches local minima of a (nonconvex) objective function which is supposed to coincide with a sum of local utility functions of the agents. The algorithm under study consists of two steps: a local stochastic gradient descent at each agent and a gossip step that drives the network of agents to a consensus. It is proved that i) an agreement is achieved between agents on the value of the estimate, ii) the algorithm converges to the set of Kuhn-Tucker points of the optimization problem. The proof relies on recent results about differential inclusions. In the context of unconstrained optimization, intelligible sufficient conditions are provided in order to ensure the stability of the algorithm. In the latter case, we also provide a central limit theorem which governs the asymptotic fluctuations of the estimate. We illustrate our results in the case of distributed power allocation for ad-hoc wireless networks.