Security-control methods for statistical databases: a comparative study
ACM Computing Surveys (CSUR)
Instance-Based Learning Algorithms
Machine Learning
On the Optimality of the Simple Bayesian Classifier under Zero-One Loss
Machine Learning - Special issue on learning with probabilistic representations
Machine Learning
ICANN '97 Proceedings of the 7th International Conference on Artificial Neural Networks
Minimum Spanning Tree Partitioning Algorithm for Microaggregation
IEEE Transactions on Knowledge and Data Engineering
Probabilistic Information Loss Measures in Confidentiality Protection of Continuous Microdata
Data Mining and Knowledge Discovery
IEEE Transactions on Knowledge and Data Engineering
\ell -Diversity: Privacy Beyond \kappa -Anonymity
ICDE '06 Proceedings of the 22nd International Conference on Data Engineering
Privacy Protection: p-Sensitive k-Anonymity Property
ICDEW '06 Proceedings of the 22nd International Conference on Data Engineering Workshops
Data Mining: Practical Machine Learning Tools and Techniques, Second Edition (Morgan Kaufmann Series in Data Management Systems)
Rethinking rank swapping to decrease disclosure risk
Data & Knowledge Engineering
Record linkage for database integration using fuzzy integrals
International Journal of Intelligent Systems
Hi-index | 0.00 |
Statistical Disclosure Control protection methods perturb the nonconfidential attributes of an original dataset and publish the perturbed results along with the values of confidential attributes. Traditionally, such a method is considered to achieve a good privacy level if attackers who try to link an original record with its perturbed counterpart have a low success probability. Another opinion is lately gaining popularity: the protection methods should resist not only record re-identification attacks, but also attacks that try to guess the true value of some confidential attribute of some original record(s). This is known as attribute disclosure risk. In this paper we propose a quite simple strategy to estimate the attribute disclosure risk suffered by a protection method: using a classifier, constructed from the protected (public) dataset, to predict the attribute values of some original record. After defining this approach in detail, we describe some experiments that show the power and danger of the approach: very popular protection methods suffer from very high attribute disclosure risk values.