Publishing anonymous survey rating data

  • Authors:
  • Xiaoxun Sun;Hua Wang;Jiuyong Li;Jian Pei

  • Affiliations:
  • Australian Council for Educational Research, Camberwell, Australia;Department of Mathematics Computing, University of Southern Queensland, Toowoomba, Australia;School of Computer and Information Science, University of South Australia, Adelaide, Australia;School of Computing Science, Simon Fraser University, Burnaby, Canada

  • Venue:
  • Data Mining and Knowledge Discovery
  • Year:
  • 2011

Quantified Score

Hi-index 0.00

Visualization

Abstract

We study the challenges of protecting privacy of individuals in the large public survey rating data in this paper. Recent study shows that personal information in supposedly anonymous movie rating records are de-identified. The survey rating data usually contains both ratings of sensitive and non-sensitive issues. The ratings of sensitive issues involve personal privacy. Even though the survey participants do not reveal any of their ratings, their survey records are potentially identifiable by using information from other public sources. None of the existing anonymisation principles (e.g., k-anonymity, l-diversity, etc.) can effectively prevent such breaches in large survey rating data sets. We tackle the problem by defining a principle called $${(k,\epsilon)}$$ -anonymity model to protect privacy. Intuitively, the principle requires that, for each transaction t in the given survey rating data T, at least (k 驴 1) other transactions in T must have ratings similar to t, where the similarity is controlled by $${\epsilon}$$ . The $${(k,\epsilon)}$$ -anonymity model is formulated by its graphical representation and a specific graph-anonymisation problem is studied by adopting graph modification with graph theory. Various cases are analyzed and methods are developed to make the updated graph meet $${(k,\epsilon)}$$ requirements. The methods are applied to two real-life data sets to demonstrate their efficiency and practical utility.