12/07/2020

Differentially Private Set Union

Pankaj Gulhane, Sivakanth Gopi, Janardhan Kulkarni, Judy Hanwen Shen, Milad Shokouhi, Sergey Yekhanin

Keywords: Privacy-preserving Statistics and Machine Learning

Abstract: We study the basic operation of set union in the global model of differential privacy. In this problem, we are given a universe $U$ of items, possibly of infinite size, and a database $D$ of users. Each user $i$ contributes a subset $W_i \subseteq U$ of items. We want an ($\epsilon$,$\delta$)-differentially private Algorithm which outputs a subset $S \subset \cup_i W_i$ such that the size of $S$ is as large as possible. The problem arises in countless real world applications, and is particularly important and ubiquitous in natural language processing (NLP) problems. For example, discovering words, sentences, $n$-grams etc., from private text data belonging to users is an instance of the set union problem. Known algorithms for this problem proceed by collecting a (weighted) subset of items from each user, taking the union of such subsets, and disclosing the items whose noisy counts fall above a certain cutoff threshold. Crucially, in the above process, the contribution of each individual user is always independent from identity of items held by other users, resulting in a wasteful aggregation process, where some items’ counts happen to be very large – far above the cutoff threshold. We deviate from the above paradigm, by allowing users to contribute their items in a {\em dependent fashion}, guided by a policy. In this new setting ensuring privacy is significantly delicate. We prove that any policy which has certain {\em contractive} properties would result in a differentially private algorithm. We design two new algorithms, one using Laplace Noise and other Gaussian noise, as specific instances of policies satisfying the contractive properties. Our experiments show that the new algorithms significantly outperform previously known mechanisms for the problem.

 0
 0
 0
 0
This is an embedded video. Talk and the respective paper are published at ICML 2020 virtual conference. If you are one of the authors of the paper and want to manage your upload, see the question "My papertalk has been externally embedded..." in the FAQ section.

Comments

Post Comment
no comments yet
code of conduct: tbd Characters remaining: 140

Similar Papers