Transparent anonymization: Thwarting adversaries who know the algorithm

  • Authors:
  • Xiaokui Xiao;Yufei Tao;Nick Koudas

  • Affiliations:
  • Nanyang Technological University, Singapore;Chinese University of Hong Kong, Shatin, Hong Kong;University of Toronto, Toronto, Canada

  • Venue:
  • ACM Transactions on Database Systems (TODS)
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

Numerous generalization techniques have been proposed for privacy-preserving data publishing. Most existing techniques, however, implicitly assume that the adversary knows little about the anonymization algorithm adopted by the data publisher. Consequently, they cannot guard against privacy attacks that exploit various characteristics of the anonymization mechanism. This article provides a practical solution tothis problem. First, we propose an analytical model for evaluating disclosure risks, when an adversary knows everything in the anonymization process, except the sensitive values. Based on this model, we develop a privacy principle, transparent l-diversity, which ensures privacy protection against such powerful adversaries. We identify three algorithms that achieve transparent l-diversity, and verify their effectiveness and efficiency through extensive experiments with real data.