Tight Differential Privacy Guarantees for the Shuffle Model with k-Randomized Response
Most differentially private algorithms assume a central model in which a reliable third party inserts noise to queries made on datasets, or a local model where the data owners directly perturb their data. However, the central model is vulnerable via a single point of failure, and the local model has the disadvantage that the utility of the data deteriorates significantly. The recently proposed shuffle model is an intermediate framework between the central and local paradigms. In the shuffle model, data owners send their locally privatized data to a server where messages are shuffled randomly, making it impossible to trace the link between a privatized message and the corresponding sender. In this paper, we theoretically derive the tightest known differential privacy guarantee for the shuffle models with k-Randomized Response (k-RR) local randomizers, under histogram queries, and we denoise the histogram produced by the shuffle model using the matrix inversion method to evaluate the utility of the privacy mechanism. We perform experiments on both synthetic and real data to compare the privacy-utility trade-off of the shuffle model with that of the central one privatized by adding the state-of-the-art Gaussian noise to each bin. We see that the difference in statistical utilities between the central and the shuffle models shows that they are almost comparable under the same level of differential privacy protection.
WOS:001280331400028
École Polytechnique Fédérale de Lausanne
Inria
Institut Polytechnique de Paris
2024-01-01
Cham
978-3-031-57536-5
978-3-031-57537-2
Lecture Notes in Computer Science; 14551
0302-9743
1611-3349
440
458
REVIEWED
EPFL
| Event name | Event acronym | Event place | Event date |
Bordeaux, FRANCE | 2023-12-11 - 2023-12-13 | ||