Experimental analysis of privacy loss in DCOP algorithms

  • Authors:
  • Rachel Greenstadt;Jonathan P. Pearce;Emma Bowring;Milind Tambe

  • Affiliations:
  • Harvard University;University of Southern California;University of Southern California;University of Southern California

  • Venue:
  • AAMAS '06 Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
  • Year:
  • 2006

Quantified Score

Hi-index 0.00

Visualization

Abstract

Distributed Constraint Optimization (DCOP) is rapidly emerging as a prominent technique for multiagent coordination. Unfortunately, rigorous quantitative evaluations of privacy loss in DCOP algorithms have been lacking despite the fact that agent privacy is a key motivation for applying DCOPs in many applications. Recently, Maheswaran et al. [3, 4] introduced a framework for quantitative evaluations of privacy in DCOP algorithms, showing that early DCOP algorithms lose more privacy than purely centralized approaches and questioning the motivation for applying DCOPs. Do state-of-the art DCOP algorithms suffer from a similar shortcoming? This paper answers that question by investigating the most efficient DCOP algorithms, including both DPOP and ADOPT.