Denials leak information: Simulatable auditing

  • Authors:
  • Krishnaram Kenthapadi;Nina Mishra;Kobbi Nissim

  • Affiliations:
  • Search Labs, Microsoft Research, Mountain View, CA 94043, United States;Search Labs, Microsoft Research, Mountain View, CA 94043, United States;Department of Computer Science, Ben-Gurion University, P.O. Box 653, Beer Sheva 84105, Israel

  • Venue:
  • Journal of Computer and System Sciences
  • Year:
  • 2013

Quantified Score

Hi-index 0.00

Visualization

Abstract

Imagine a data set consisting of private information about individuals. The online query auditing problem is: given a sequence of queries that have already been posed about the data, their corresponding answers and given a new query, deny the answer if privacy can be breached or give the true answer otherwise. We investigate the fundamental problem that query denials leak information. This problem was largely overlooked in previous work on auditing. Because of this oversight, some of the previously suggested auditors can be used by an attacker to compromise the privacy of a large fraction of the individuals in the data. To overcome this problem, we introduce a new model called simulatable auditing where query denials provably do not leak information. We present a simulatable auditing algorithm for max queries under the classical definition of privacy where a breach occurs if a sensitive value is fully compromised. Because of the known limitations of the classical definition of compromise, we describe a probabilistic notion of (partial) compromise, closely related to the notion of semantic security. We demonstrate that sum queries can be audited in a simulatable fashion under probabilistic compromise, making some distributional assumptions.