This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Provable Membership Inference Privacy

Zachary Izzo Department of Mathematics, Stanford University, zizzo@stanford.edu. Research was conducted while the author was an intern at Google Cloud AI Research.    Jinsung Yoon Google Cloud AI Research, jinsungyoon@google.com    Sercan O. Arik Google Cloud AI Research, soarik@google.com    James Zou Department of Biomedical Data Science, Stanford University, jamesz@stanford.edu
Abstract

In applications involving sensitive data, such as finance and healthcare, the necessity for preserving data privacy can be a significant barrier to machine learning model development. Differential privacy (DP) has emerged as one canonical standard for provable privacy. However, DP’s strong theoretical guarantees often come at the cost of a large drop in its utility for machine learning; and DP guarantees themselves can be difficult to interpret. In this work, we propose a novel privacy notion, membership inference privacy (MIP), to address these challenges. We give a precise characterization of the relationship between MIP and DP, and show that MIP can be achieved using less amount of randomness compared to the amount required for guaranteeing DP, leading to smaller drop in utility. MIP guarantees are also easily interpretable in terms of the success rate of membership inference attacks. Our theoretical results also give rise to a simple algorithm for guaranteeing MIP which can be used as a wrapper around any algorithm with a continuous output, including parametric model training.

1 Introduction

As the popularity and efficacy of machine learning (ML) have increased, the number of domains in which ML is applied has also expanded greatly. Some of these domains, such as finance or healthcare, have ML on sensitive data which cannot be publicly shared due to regulatory or ethical concerns (Assefa et al., 2020; Office for Civil Rights, 2002). In these instances, maintaining data privacy is of paramount importance and must be considered at every stage of the ML process, from model development to deployment. During development, even sharing data in-house while retaining the appropriate level of privacy can be a barrier to model development (Assefa et al., 2020). After deployment, the trained model itself can leak information about the training data if appropriate precautions are not taken (Shokri et al., 2017; Carlini et al., 2021a).

Differential privacy (DP) (Dwork et al., 2014) has emerged as the gold standard for provable privacy in the academic literature. Training methods for DP use randomized algorithms applied on databases of points, and DP stipulates that the algorithm’s random output cannot change much depending on the presence or absence of one individual point in the database. These guarantees in turn give information theoretic protection against the maximum amount of information that an adversary can obtain about any particular sample in the dataset, regardless of that adversary’s prior knowledge or computational power, making DP an attractive method for guaranteeing privacy. However, DP’s strong theoretical guarantees often come at the cost of a large drop in utility for many algorithms (Jayaraman and Evans, 2019). In addition, DP guarantees themselves are difficult to interpret by non-experts. There is a precise definition for what it means for an algorithm to satisfy DP with ε=10\varepsilon=10, but it is not a priori clear what this definition guarantees in terms of practical questions that a user could have, the most basic of which might be to ask whether or not an attacker can determine whether or not that user’s information was included in the algorithm’s input. These constitute challenges for adoption of DP in practice.

In this paper, we propose a novel privacy notion, membership inference privacy (MIP), to address these challenges. Membership inference measures privacy via a game played between the algorithm designer and an adversary or attacker. The adversary is presented with the algorithm’s output and a “target” sample, which may or may not have been included in the algorithm’s input set. The adversary’s goal is to determine whether or not the target sample was included in the algorithm’s input. If the adversary can succeed with probability much higher than random guessing, then the algorithm must be leaking information about its input. This measure of privacy is one of the simplest for the attacker; thus, provably protecting against it is a strong privacy guarantee. Furthermore, MIP is easily interpretable, as it is measured with respect to a simple quantity–namely, the maximum success rate of an attacker. In summary, our contributions are as follows:

  • We propose a novel privacy notion, which we dub membership inference privacy (MIP).

  • We characterize the relationship between MIP and differential privacy (DP). In particular, we show that DP is sufficient to certify MIP and quantify the correspondence.

  • In addition, we demonstrate that in some cases, MIP can be certified using less randomness than that required for certifying DP. (In other words, while DP is sufficient for certifying MIP, it is not necessary.) This in turn generally means that MIP algorithms can have greater utility than those which implement DP.

  • We introduce a “wrapper” method for turning any base algorithm with continuous output into an algorithm which satisfies MIP.

2 Related Work

Privacy Attacks in ML

The study of privacy attacks has recently gained popularity in the machine learning community as the importance of data privacy has become more apparent. In a membership inference attack (Shokri et al., 2017), an attacker is presented with a particular sample and the output of the algorithm to be attacked. The attacker’s goal is to determine whether or not the presented sample was included in the training data or not. If the attacker can determine the membership of the sample with a probability significantly greater than random guessing, this indicates that the algorithm is leaking information about its training data. Obscuring whether or not a given individual belongs to the private dataset is the core promise of private data sharing, and the main reason that we focus on membership inference as the privacy measure. Membership inference attacks against predictive models have been studied extensively (Shokri et al., 2017; Baluta et al., 2022; Hu et al., 2022; Liu et al., 2022; He et al., 2022; Carlini et al., 2021a), and recent work has also developed membership inference attacks against synthetic data (Stadler et al., 2022; Chen et al., 2020).

In a reconstruction attack, the attacker is not presented with a real sample to classify as belonging to the training set or not, but rather has to create samples belonging to the training set based only on the algorithm’s output. Reconstruction attacks have been successfully conducted against large language models (Carlini et al., 2021b). At present, these attacks require the attacker to have a great deal of auxiliary information to succeed. For our purposes, we are interested in privacy attacks to measure the privacy of an algorithm, and such a granular task may place too high burden on the attacker to accurately detect “small” amounts of privacy leakage.

In an attribute inference attack (Bun et al., 2021; Stadler et al., 2022), the attacker tries to infer a sensitive attribute from a particular sample, based on its non-sensitive attributes and the attacked algorithm output. It has been argued that attribute inference is really the entire goal of statistical learning, and therefore should not be considered a privacy violation (Bun et al., 2021; Jayaraman and Evans, 2022).

Differential Privacy (DP)

DP (Dwork et al., 2014) and its variants (Mironov, 2017; Dwork and Rothblum, 2016) offer strong, information-theoretic privacy guarantees. A DP (probabilistic) algorithm is one in which the probability law of its output does not change much if one sample in its input is changed. That is, if DD and DD^{\prime} are two adjacent datasets (i.e., two datasets which differ in exactly one element), then the algorithm 𝒜\mathcal{A} is ε\varepsilon-DP if (𝒜(D)S)eε(𝒜(D)S)\mathbb{P}(\mathcal{A}(D)\in S)\leq e^{\varepsilon}\mathbb{P}(\mathcal{A}(D^{\prime})\in S) for any subset SS of the output space. DP has many desirable properties, such as the ability to compose DP methods or post-process the output without losing guarantees. Many simple “wrapper” methods are also available for certifying DP. Among the simplest, the Laplace mechanism, adds Laplace noise to the algorithm output. The noise level must generally depend on the sensitivity of the base algorithm, which measures how much a single input sample can change the algorithm’s output. The method we propose in this work is similar to the Laplace mechanism, but we show that the amount of noise needed can be reduced drastically. Abadi et al. (2016) introduced DP-SGD, a powerful tool enabling DP to be combined with deep learning, based on a small modification to the standard gradient descent training. However, enforcing DP does not come without a cost. Enforcing DP with high levels of privacy (small ε\varepsilon) often comes with sharp decreases in algorithm utility (Tao et al., 2021; Stadler et al., 2022). DP is also difficult to audit; it must be proven mathematically for a given algorithm. Checking it empirically is generally computationally intractable (Gilbert and McMillan, 2018). The difficulty of checking DP has led to widespread implementation issues (and even errors due to finite machine precision), which invalidate the guarantees of DP (Jagielski et al., 2020).

While the basic definition of DP can be difficult to interpret, equivalent “operational” definitions have been developed (Wasserman and Zhou, 2010; Kairouz et al., 2015; Nasr et al., 2021). These works show that DP can equivalently be expressed in terms of the maximum success rate on an adversary which seeks to distinguish between two adjacent datasets DD and DD^{\prime}, given only the output of a DP algorithm. While similar to the setting of membership inference at face value, there are subtle differences. In particular, in the case of membership inference, one must consider all datasets which could have contained the target record, and all datasets which do not contain the target record, and distinguish between the algorithm’s output in this larger space of possibilities.

Lastly, the independent work of Thudi et al. (2022) specifically applies DP to bound membership inference rates, and our results in Sec. 3.4 complement theirs on the relationship between membership inference and DP. However, our results show that DP is not required to prevent membership inference; it is merely one option, and we give alternative methods for defending against membership inference.

3 Membership Inference Privacy

In this section, we will motivate and define membership inference privacy and derive several theoretical results pertaining to it. We will provide proof sketches for many of our results, but the complete proofs all propositions, theorems, etc., can be found in Appendix A.

3.1 Notation

We make use of the following notation. We will use 𝒟\mathcal{D} to refer to our entire dataset, which consists of nn samples all of which must remain private. We will use 𝐱𝒟\mathbf{x}\in\mathcal{D} or 𝐱𝒟\mathbf{x}^{*}\in\mathcal{D} to refer to a particular sample. 𝒟train𝒟\mathcal{D}_{\mathrm{train}}\subseteq\mathcal{D} refers to a size-kk subset of 𝒟\mathcal{D}. We will assume the subset is selected randomly, so 𝒟train\mathcal{D}_{\mathrm{train}} is a random variable. The remaining data 𝒟𝒟train\mathcal{D}\setminus\mathcal{D}_{\mathrm{train}} will be referred to as the holdout data. We denote by 𝔻\mathbb{D} the set of all size-kk subsets of 𝒟\mathcal{D} (i.e., all possible training sets), and we will use D𝔻D\in\mathbb{D} to refer to a particular realization of the random variable 𝒟train\mathcal{D}_{\mathrm{train}}. Finally, given a particular sample 𝐱𝒟\mathbf{x}^{*}\in\mathcal{D}, 𝔻in\mathbb{D}^{\mathrm{in}} (resp. 𝔻out\mathbb{D}^{\mathrm{out}}) will refer to sets D𝔻D\in\mathbb{D} for which 𝐱D\mathbf{x}^{*}\in D (resp. 𝐱D\mathbf{x}^{*}\not\in D).

3.2 Theoretical Motivation

The implicit assumption behind the public release of any statistical algorithm–be it a generative or predictive ML model, or even the release of simple population statistics–is that it is acceptable for statistical information about the modelled data to be released publicly. In the context of membership inference, this poses a potential problem: if the population we are modeling is significantly different from the “larger” population, then if our algorithm’s output contains any useful information whatsoever, it should be possible for an attacker to infer whether or not a given record could have plausibly come from our training data or not.

We illustrate this concept with an example. Suppose the goal is to publish an ML model which predicts a patient’s blood pressure from several biomarkers, specifically for patients who suffer from a particular chronic disease. To do this, we collect a dataset of individuals with confirmed cases of the disease, and use this data to train a linear regression model with coefficients θ^\hat{\theta}. Formally, we let 𝐱d\mathbf{x}\in\mathbb{R}^{d} denote the features (e.g. biomarker values), zz\in\mathbb{R} denote the patient’s blood pressure, and y=𝟙{patient has the chronic disease in question}y=\mathds{1}\{\textrm{patient has the chronic disease in question}\}. In this case, the private dataset 𝒟train\mathcal{D}_{\mathrm{train}} contains only the patients with y=1y=1. Assume that in the general populace, patient features are drawn from a mixture model:

yBernoulli(p),𝐱𝒩(0,I),z|𝐱,yθy𝐱,θ0θ1.\displaystyle y\sim\mathrm{Bernoulli}(p),\hskip 18.06749pt\mathbf{x}\sim\mathcal{N}(0,I),\hskip 18.06749ptz|\mathbf{x},y\sim\theta_{y}^{\top}\mathbf{x},\hskip 18.06749pt\theta_{0}\neq\theta_{1}.

In the membership inference attack scenario, an adversary observes a data point (𝐱,z)(\mathbf{x}^{*},z^{*}) and the model θ^\hat{\theta}, and tries to determine whether or not (𝐱,z)𝒟train(\mathbf{x}^{*},z^{*})\in\mathcal{D}_{\mathrm{train}}. If θ0\theta_{0} and θ1\theta_{1} are well-separated, then an adversary can train an effective classifier to determine the corresponding label 𝟙{(𝐱,z)𝒟train}\mathds{1}\{(\mathbf{x}^{*},z^{*})\in\mathcal{D}_{\mathrm{train}}\} for (𝐱,z)(\mathbf{x}^{*},z^{*}) by checking whether or not zθ^𝐱z^{*}\approx\hat{\theta}^{\top}\mathbf{x}^{*}. Since only data with y=1y=1 belong to 𝒟train\mathcal{D}_{\mathrm{train}}, this provides a signal to the adversary as to whether or not 𝐱\mathbf{x}^{*} could have belonged to 𝒟train\mathcal{D}_{\mathrm{train}} or not. The point is that in this setting, this outcome is unavoidable if θ^\hat{\theta} is to provide any utility whatsoever. In other words:

In order to preserve utility, membership inference privacy must be measured with respect to the distribution from which the private data are drawn.

The example above motivates the following theoretical ideal for our membership inference setting. Let 𝒟={𝐱i}i=1n\mathcal{D}=\{\mathbf{x}_{i}\}_{i=1}^{n} be the private dataset and suppose that 𝐱ii.i.d.𝒫\mathbf{x}_{i}\stackrel{{\scriptstyle\tiny{\mathrm{i.i.d.}}}}{{\sim}}\mathcal{P} for some probability distribution 𝒫\mathcal{P}. (Note: Here, 𝐱\mathbf{x}^{*} corresponds to the complete datapoint (𝐱,z)(\mathbf{x}^{*},z^{*}) in the example above.) Let 𝒜\mathcal{A} be our (randomized) algorithm, and denote its output by θ=𝒜(𝒟)\theta=\mathcal{A}(\mathcal{D}). We generate a test point based on:

yBernoulli(1/2),𝐱|yyUnif(𝒟train)+(1y)𝒫,y^{*}\sim\mathrm{Bernoulli}\left(1/2\right),\hskip 18.06749pt\mathbf{x}^{*}|y^{*}\sim y^{*}\textrm{Unif}(\mathcal{D}_{\mathrm{train}})+(1-y^{*})\mathcal{P},

i.e. 𝐱\mathbf{x}^{*} is a fresh draw from 𝒫\mathcal{P} or a random element of the private training data with equal probability. Let \mathcal{I} denote any membership inference algorithm which takes as input 𝐱\mathbf{x}^{*} and the algorithm’s output θ=𝒜(𝒟train)\theta=\mathcal{A}(\mathcal{D}_{\mathrm{train}}). The notion of privacy we wish to enforce is that \mathcal{I} cannot do much better to ascertain the membership of 𝐱\mathbf{x}^{*} than guessing randomly:

𝒜,𝒟train((𝐱,θ)=y)1/2+η,\mathbb{P}_{\mathcal{A},\mathcal{D}_{\mathrm{train}}}(\mathcal{I}(\mathbf{x}^{*},\theta)=y^{*})\leq 1/2+\eta, (1)

where ideally η1/2\eta\ll 1/2.

3.3 Practical Definition

In reality, we do not have access to the underlying distribution 𝒫\mathcal{P}. Instead, we propose to use a bootstrap sampling approach to approximate fresh draws from 𝒫\mathcal{P}.

Definition 1 (Membership Inference Privacy (MIP)).

Given fixed knk\leq n, let 𝒟train𝒟\mathcal{D}_{\mathrm{train}}\subseteq\mathcal{D} be a size-kk subset chosen uniformly at random from the elements in 𝒟\mathcal{D}. For 𝐱𝒟\mathbf{x}^{*}\in\mathcal{D}, let y=𝟙{𝐱𝒟train}y^{*}=\mathds{1}\{\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\}. An algorithm 𝒜\mathcal{A} is η\eta-MIP with respect to 𝒟\mathcal{D} if for any identification algorithm \mathcal{I} and for every 𝐱𝒟\mathbf{x}^{*}\in\mathcal{D}, we have

((𝐱,𝒜(𝒟train))=y)max{kn,1kn}+η.\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(\mathcal{D}_{\mathrm{train}}))=y^{*})\leq\max\left\{\frac{k}{n},1-\frac{k}{n}\right\}+\eta.

Here, the probability is taken over the uniformly random size-kk subset 𝒟train𝒟\mathcal{D}_{\mathrm{train}}\subseteq\mathcal{D}, as well as any randomness in 𝒜\mathcal{A} and \mathcal{I}.

Definition 1 states that given the output of 𝒜\mathcal{A}, an adversary cannot determine whether a given point was in the holdout set or training set with probability more than η\eta better than always guessing the a priori more likely outcome. In the remainder of the paper, we will set k=n/2k=n/2, so that 𝒜\mathcal{A} is η\eta-MIP if an attacker cannot have average accuracy greater than (1/2+η)(1/2+\eta). This gives the largest a priori entropy for the attacker’s classification task, which creates the highest ceiling on how much of an advantage an attacker can possibly gain from the algorithm’s output, and consequently the most accurate measurement of privacy leakage. The choice k=n/2k=n/2 also keeps us as close as possible to the theoretical motivation in the previous subsection. We note that analogues of all of our results apply for general kk.

The definition of MIP is phrased with respect to any classifier (whose randomness is independent of the randomness in 𝒜\mathcal{A}; if the adversary knows the algorithm and the random seed, we are doomed). While this definition is compelling in that it shows a bound on what any attacker can hope to accomplish, the need to consider all possible attack algorithms makes it difficult to work with technically. The following proposition shows that MIP is equivalent to a simpler definition which does not need to simultaneously consider all identification algorithms \mathcal{I}.

Proposition 2.

Let 𝔸=Range(𝒜)\mathbb{A}=\mathrm{Range}(\mathcal{A}) and let μ\mu denote the probability law of 𝒜(𝒟train)\mathcal{A}(\mathcal{D}_{\mathrm{train}}). Then 𝒜\mathcal{A} is η\eta-MIP if and only if

𝔸(max{(𝐱𝒟train|𝒜(𝒟train)=A),(𝐱𝒟train|𝒜(𝒟train)=A)}dμ(A))12+η.\displaystyle\int_{\mathbb{A}}\Big{(}\max\big{\{}\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}})=A),\mathbb{P}(\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}})=A)\big{\}}\,d\mu(A)\Big{)}\leq\frac{1}{2}+\eta.

Furthermore, the optimal adversary is given by

(𝐱,A)=𝟙{(𝐱𝒟train|𝒜(𝒟train)=A)1/2}.\mathcal{I}(\mathbf{x}^{*},A)=\mathds{1}\{\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}})=A)\geq 1/2\}.

Proposition 2 makes precise the intuition that the optimal attacker should guess the more likely of 𝐱𝒟train\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}} or 𝐱𝒟train\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}} conditional on the output of 𝒜\mathcal{A}. The optimal attacker’s overall accuracy is then computed by marginalizing this conditional statement.

Finally, MIP also satisfies a post-processing inequality similar to the classical result in DP (Dwork et al., 2014). This states that any local functions of a MIP algorithm’s output cannot degrade the privacy guarantee.

Theorem 3.

Suppose that 𝒜\mathcal{A} is η\eta-MIP, and let ff be any (potentially randomized, with randomness independent of 𝒟train\mathcal{D}_{\mathrm{train}}) function. Then f𝒜f\circ\mathcal{A} is also η\eta-MIP.

Proof.

Let f\mathcal{I}_{f} be any membership inference algorithm for f𝒜f\circ\mathcal{A}. Define 𝒜(𝐱,𝒜(𝒟train))=f(𝐱,f(𝒜(𝒟train)))\mathcal{I}_{\mathcal{A}}(\mathbf{x}^{*},\mathcal{A}(\mathcal{D}_{\mathrm{train}}))=\mathcal{I}_{f}(\mathbf{x}^{*},f(\mathcal{A}(\mathcal{D}_{\mathrm{train}}))). Since 𝒜\mathcal{A} is η\eta-MIP, we have

12+η(𝒜(𝐱,𝒜(𝒟train))=y)=(f(𝐱,f(𝒜(𝒟train)))=y).\displaystyle\frac{1}{2}+\eta\geq\mathbb{P}(\mathcal{I}_{\mathcal{A}}(\mathbf{x}^{*},\mathcal{A}(\mathcal{D}_{\mathrm{train}}))=y^{*})=\mathbb{P}(\mathcal{I}_{f}(\mathbf{x}^{*},f(\mathcal{A}(\mathcal{D}_{\mathrm{train}})))=y^{*}).

Thus, f𝒜f\circ\mathcal{A} is η\eta-MIP by Definition 1. ∎

For example, Theorem 3 is important for the application of MIP to generative model training – if we can guarantee that our generative model is η\eta-MIP, then any output produced by it is η\eta-MIP as well.

3.4 Relation to Differential Privacy

In this section, we make precise the relationship between MIP and the most common theoretical formulation of privacy: differential privacy (DP). We provide proof sketches for most of our results here; detailed proofs can be found in the Appendix. Our first theorem shows that DP is at least as strong as MIP.

Theorem 4.

Let 𝒜\mathcal{A} be ε\varepsilon-DP. Then 𝒜\mathcal{A} is η\eta-MIP with η=11+eε12\eta=\frac{1}{1+e^{-\varepsilon}}-\frac{1}{2}. Furthermore, this bound is tight, i.e. for any ε>0\varepsilon>0, there exists an ε\varepsilon-DP algorithm against which the optimal attacker has accuracy 11+eε\frac{1}{1+e^{-\varepsilon}}.

Proof sketch.

Let p=(𝐱𝒟train|𝒜(𝒟train))p=\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}})) and q=(𝐱𝒟train|𝒜(𝒟train))q=\mathbb{P}(\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}})) and suppose without loss of generality that qpq\geq p. We have p+q=1p+q=1 and by Proposition 6 below, q/peεq/p\leq e^{\varepsilon}. This implies that q11+eεq\leq\frac{1}{1+e^{-\varepsilon}}, and applying Proposition 2 gives the desired result.

For the tightness result, there is a simple construction on subsets of size 1 of 𝒟={0,1}\mathcal{D}=\{0,1\}. Let p=11+eεp=\frac{1}{1+e^{-\varepsilon}} and q=1pq=1-p. The algorithm 𝒜(D)\mathcal{A}(D) which outputs DD with probability pp and 𝒟D\mathcal{D}\setminus D with probability qq is ε\varepsilon-DP, and the optimal attacker has exactly the accuracy given in the theorem. ∎

To help interpret this result, we remark that for ε0\varepsilon\approx 0, we have 11+eε12ε/4\frac{1}{1+e^{-\varepsilon}}-\frac{1}{2}\approx\varepsilon/4. Thus in the regime where strong privacy guarantees are required (η0\eta\approx 0), ηε/4\eta\approx\varepsilon/4.

In fact, DP is strictly stronger than MIP, which we make precise with the following theorem.

Theorem 5.

For any η>0\eta>0, there exists an algorithm 𝒜\mathcal{A} which is η\eta-MIP but not ε\varepsilon-DP for any ε<\varepsilon<\infty.

Proof sketch.

The easiest example is an algorithm which publishes each sample in its input set with extremely low probability. Since the probability that any given sample is published is low, the probability that an attacker can do better than guess randomly is low marginally over the algorithm’s output. However, adding a sample to the input dataset changes the probability of that sample’s being published from 0 to a strictly positive number, so the guarantee on probability ratios required for DP is infinite. ∎

In order to better understand the difference between DP and MIP, let us again examine Proposition 2. Recall that this proposition showed that marginally over the output of 𝒜\mathcal{A}, the conditional probability that 𝐱𝒟train\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}} given the algorithm output should not differ too much from the unconditional probability that 𝐱𝒟train\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}. The following proposition shows that DP requires this condition to hold for every output of 𝒜(𝒟train)\mathcal{A}(\mathcal{D}_{\mathrm{train}}).

Proposition 6.

If 𝒜\mathcal{A} is an ε\varepsilon-DP algorithm, then for any 𝐱\mathbf{x}^{*}, we have

(𝐱𝒟train|𝒜(𝒟train))(𝐱𝒟train|𝒜(𝒟train))eε(𝐱𝒟train)(𝐱𝒟train).\frac{\mathbb{P}(\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}}))}{\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}}))}\leq e^{\varepsilon}\frac{\mathbb{P}(\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}})}{\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}})}.

Proposition 6 can be thought of as an extension of the Bayesian interpretation of DP explained by Jordon et al. (2022). Namely, the definition of DP immediately implies that, for any two adjacent sets DD and DD^{\prime},

(𝒟train=D|𝒜(𝒟train))(𝒟train=D|𝒜(𝒟train))eε(𝒟train=D)(𝒟train=D).\frac{\mathbb{P}(\mathcal{D}_{\mathrm{train}}=D\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}}))}{\mathbb{P}(\mathcal{D}_{\mathrm{train}}=D^{\prime}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}}))}\leq e^{\varepsilon}\frac{\mathbb{P}(\mathcal{D}_{\mathrm{train}}=D)}{\mathbb{P}(\mathcal{D}_{\mathrm{train}}=D^{\prime})}.

We remark that the proof of Proposition 6 indicates that converting between the case of distinguishing between two adjacent datasets (as in the inequality above, and as done in (Wasserman and Zhou, 2010; Kairouz et al., 2015; Nasr et al., 2021)) vs. the case of membership inference is non-trivial: both our proof and a similar one by Thudi et al. (2022) require the construction of a injective function between sets which do/do not contain 𝐱\mathbf{x}^{*}.

4 Guaranteeing MIP via Noise Addition

In this section, we show that a small modification to standard training procedures can be used to guarantee MIP. Suppose that 𝒜\mathcal{A} takes as input a data set DD and produces output θd\theta\in\mathbb{R}^{d}. For instance, 𝒜\mathcal{A} may compute a simple statistical query on DD, such as mean estimation, but our results apply equally well in the case that e.g. 𝒜(D)\mathcal{A}(D) are the weights of a neural network trained on DD. If θ\theta are the weights of a generative model, then if we can guarantee MIP for θ\theta, then by the data processing inequality (Theorem 3), this guarantees privacy for any output of the generative model.

The distribution over training data (in our case, the uniform distribution over size n/2n/2 subsets of our complete dataset 𝒟\mathcal{D}) induces a distribution over the output θ\theta. The question is, what is the smallest amount of noise we can add to θ\theta which will guarantee MIP? If we add noise on the order of maxDD𝒟𝒜(D)𝒜(D)\max_{D\sim D^{\prime}\subseteq\mathcal{D}}\|\mathcal{A}(D)-\mathcal{A}(D^{\prime})\|, then we can adapt the standard proof for guaranteeing DP in terms of algorithm sensitivity to show that a restricted version of DP (only with respect subsets of 𝒟\mathcal{D}) holds in this case, which in turn guarantees MIP. However, recall that by Propositions 2 and 6, MIP is only asking for a marginal guarantee on the change in the posterior probability of DD given AA, whereas DP is asking for a conditional guarantee on the posterior. So while max\max seems necessary for a conditional guarantee, the moments of θ\theta should be sufficient for a marginal guarantee. Theorem 7 shows that this intuition is correct.

Theorem 7.

Let \|\cdot\| be any norm, and let σM𝔼θ𝔼θM\sigma^{M}\geq\mathbb{E}\|\theta-\mathbb{E}\theta\|^{M} be an upper bound on the MM-th central moment of θ\theta with respect to this norm over the randomness in 𝒟train\mathcal{D}_{\mathrm{train}} and 𝒜\mathcal{A}. Let XX be a random variable with density proportional to exp(1cσX)\exp(-\frac{1}{c\sigma}\|X\|) with c=(7.5/η)1+2Mc=(7.5/\eta)^{1+\frac{2}{M}}. Finally, let θ^=θ+X\hat{\theta}=\theta+X. Then θ^\hat{\theta} is η\eta-MIP, i.e., for any adversary \mathcal{I},

((𝐱,θ^)=y)1/2+η.\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\hat{\theta})=y^{*})\leq 1/2+\eta.
Proof sketch.

The proof proceeds by bounding the posterior likelihood ratio (𝐱𝒟train|θ^)(𝐱𝒟train|θ^)\frac{\mathbb{P}(\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}}\>|\>\hat{\theta})}{\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\>|\>\hat{\theta})} from above and below for all θ^\hat{\theta} in a large \|\cdot\|-ball. This in turn yields an upper bound on the max in the integrand in Proposition 2 with high probability over 𝒜(𝒟train)\mathcal{A}(\mathcal{D}_{\mathrm{train}}). The central moment σ\sigma allows us to apply a generalized Chebyshev inequality to establish these bounds. The full proof is computationally intensive and the complete details can be found in the Appendix. ∎

At first glance, Theorem 7 may appear to be adding noise of equal magnitude to all of the coordinates of θ\theta, regardless of how much each contributes to the central moment σ\sigma. However, by carefully selecting the norm \|\cdot\|, we can add non-isotropic noise to θ\theta such that the marginal noise level reflects the variability of each specific coordinate of θ\theta. This is the content of Corollary 8. (GenNormal(μ,α,β)\mathrm{GenNormal}(\mu,\alpha,\beta) refers to the probability distribution with density proportional to exp(((xμ)/α)β)\exp(-((x-\mu)/\alpha)^{\beta}).)

Corollary 8.

Let M2M\geq 2, σiM𝔼|θi𝔼θi|M\sigma_{i}^{M}\geq\mathbb{E}|\theta_{i}-\mathbb{E}\theta_{i}|^{M}, and define xσ,M=(i=1d|xi|MdσiM)1/M\|x\|_{\sigma,M}=\left(\sum_{i=1}^{d}\frac{|x_{i}|^{M}}{d\sigma_{i}^{M}}\right)^{1/M}. Generate

YiGenNormal(0,σi,M),U=Y/Yσ,MY_{i}\sim\mathrm{GenNormal}(0,\sigma_{i},M),\hskip 10.84006ptU=Y/\|Y\|_{\sigma,M}

and draw rLaplace((6.16/η)1+2/M)r\sim\mathrm{Laplace}\left((6.16/\eta)^{1+2/M}\right). Finally, set X=rUX=rU and return θ^=θ+X\hat{\theta}=\theta+X. Then θ^\hat{\theta} is η\eta-MIP.

Proof sketch.

Let =σ,M\|\cdot\|=\|\cdot\|_{\sigma,M}. It can be shown that the density of XX has the proper form. Furthermore, by definition of the σi\sigma_{i}, we have 𝔼θ𝔼θM1\mathbb{E}\|\theta-\mathbb{E}\theta\|^{M}\leq 1. The corollary follows directly from Theorem 7. The improvement in the numerical constant (from 7.5 to 6.16) comes from numerically optimizing some of the bounds in Theorem 7, and these optimizations are valid for M2M\geq 2. ∎

Algorithm 1 MIP via noise addition
Private dataset 𝒟\mathcal{D}, σ\sigma estimation budget BB, MIP parameter η\eta
𝒟trainRandomSplit(𝒟,1/2)\mathcal{D}_{\mathrm{train}}\leftarrow\textsc{RandomSplit}(\mathcal{D},1/2)
# Estimate σ\sigma if an a priori bound is not known
for i=1,,Bi=1,\ldots,B do
     𝒟train(i)RandomSplit(𝒟train,1/2)\mathcal{D}_{\mathrm{train}}^{(i)}\leftarrow\textsc{RandomSplit}(\mathcal{D}_{\mathrm{train}},1/2)
     θ(i)𝒜(𝒟train(i))\theta^{(i)}\leftarrow\mathcal{A}(\mathcal{D}_{\mathrm{train}}^{(i)})
end for
for j=1,,dj=1,\ldots,d do
     θ¯j1Bi=1Bθj(i)\bar{\theta}_{j}\leftarrow\frac{1}{B}\sum_{i=1}^{B}\theta^{(i)}_{j}
     σj(1Bi=1B(θj(i)θ¯j)M)1/M\sigma_{j}\leftarrow\left(\frac{1}{B}\sum_{i=1}^{B}(\theta^{(i)}_{j}-\bar{\theta}_{j})^{M}\right)^{1/M}
end for
# Add appropriate noise to the base algorithm’s output
UUnif({ud:uσ,M=1})U\leftarrow\mathrm{Unif}(\{u\in\mathbb{R}^{d}\>:\>\|u\|_{\sigma,M}=1\})
rLaplace((6.16η)1+2/M)r\leftarrow\mathrm{Laplace}\left(\left(\frac{6.16}{\eta}\right)^{1+2/M}\right)
XrUX\leftarrow rU
return 𝒜(𝒟train)+X\mathcal{A}(\mathcal{D}_{\mathrm{train}})+X

In practice, the σi\sigma_{i} may not be known, but they can easily be estimated from data. We implement this intuition and devise a practical method for guaranteeing MIP in Algorithm 1. We remark briefly that the estimator for σj\sigma_{j} used in Algorithm 1 is not unbiased, but it is consistent (i.e., the bias approaches 0 as BB\rightarrow\infty). When M=2M=2, there is a well-known unbiased estimate for the variance which replace 1/B1/B with 1/(B1)1/(B-1), and one can make similar corrections for general MM (Gerlovina and Hubbard, 2019). In practice, these corrections yield very small difference and the naive estimator presented in the algorithm should suffice.

When Does MIP Improve Over DP?

By Theorem 4, any DP algorithm gives rise to a MIP algorithm, so we never need to add more noise than the amount required to guarantee DP, in order to guarantee MIP. However, Theorem 7 shows that MIP affords an advantage over DP when the variance of our algorithm’s output (over subsets of size n/2n/2) is much smaller than its sensitivity Δ\Delta, which is defined as the maximum change in the algorithm’s output when evaluated on two datasets which differ in only one element. For instance, applying the Laplace mechanism from DP requires a noise which scales like Δ/ϵ\Delta/\epsilon to guarantee ε\varepsilon-DP. It is easy to construct examples where the variance is much smaller than the sensitivity if the output of our “algorithm” is allowed to be completely arbitrary as a function of the input. However, it is more interesting to ask if there are any natural settings in which this occurs. Proposition 9 answers this question in the affirmative.

Proposition 9.

For any finite DD\subseteq\mathbb{R}, define 𝒜(D)=1xDx\mathcal{A}(D)=\frac{1}{\sum_{x\in D}x}. Given a dataset 𝒟\mathcal{D} of size nn, define 𝔻={D𝒟:|D|=n/2},\mathbb{D}=\{D\subseteq\mathcal{D}\>:\>|D|=\lfloor n/2\rfloor\}, and define

σ2=Var(𝒜(D)),Δ=maxDD𝔻|𝒜(D)𝒜(D)|.\sigma^{2}=\mathrm{Var}(\mathcal{A}(D)),\hskip 18.06749pt\Delta=\max_{D\sim D^{\prime}\in\mathbb{D}}|\mathcal{A}(D)-\mathcal{A}(D^{\prime})|.

Here the variance is taken over DUnif(𝔻)D\sim\mathrm{Unif}(\mathbb{D}). Then for all nn, there exists a dataset |𝒟|=n|\mathcal{D}|=n such that σ2=O(1)\sigma^{2}=O(1) but Δ=Ω(2n/3)\Delta=\Omega(2^{n/3}).

Proof sketch.

Assume nn is even for simplicity. Let p=(nn/2)1p=\binom{n}{n/2}^{-1} and A=pi=0n222iA=\sqrt{p}-\sum_{i=0}^{\frac{n}{2}-2}2^{i}. Take

𝒟={2i:i=0,,n2}{A}.\mathcal{D}=\{2^{i}\>:\>i=0,\ldots,n-2\}\cup\{A\}.

When D={20,,2n22,A}D=\{2^{0},\ldots,2^{\frac{n}{2}-2},A\}, then 𝒜(D)=p1/2\mathcal{A}(D)=p^{-1/2}, and this occurs with probability pp. For all other subsets DD^{\prime}, 0𝒜(D)10\leq\mathcal{A}(D^{\prime})\leq 1. ∎

5 Simulation Results

5.1 Noise Level in Proposition 9

To illustrate our theoretical results, we plot the noise level needed to guarantee MIP vs. the corresponding level of DP (with the correspondence given by Theorem 4) for the example in Proposition 9.

Refer to Fig. 1. Dotted lines refer to DP, while the solid line is for MIP with M=2M=2. The xx-axis gives the best possible bound on the attacker’s improvement in accuracy over random guessing–i.e., the parameter η\eta for an η\eta-MIP method–according to that method’s guarantees. For DP, the value along the xx-axis is given by the (tight) correspondence in Theorem 4, namely η=11+eε12\eta=\frac{1}{1+e^{-\varepsilon}}-\frac{1}{2}. η=0\eta=0 corresponds to perfect privacy (the attacker cannot do any better than random guessing), while η=12\eta=\frac{1}{2} corresponds to no privacy (the attacker can determine membership with perfect accuracy). The yy-axis denotes the amount of noise that must be added to the non-private algorithm’s output, as measured by the scale parameter of the Laplace noise that must be added (lower is better). For MIP, by Theorem 7, this is (6.16/η)2σ(6.16/\eta)^{2}\sigma where σ\sigma is an upper bound on the variance of the base algorithm over random subsets, and for DP this is Δlog1+2η12η\frac{\Delta}{\log\frac{1+2\eta}{1-2\eta}}. (This comes from solving η=11+eε12\eta=\frac{1}{1+e^{-\varepsilon}}-\frac{1}{2} for ε\varepsilon, then using the fact that Laplace(Δ/ε)\mathrm{Laplace}(\Delta/\varepsilon) noise must be added to guarantee ε\varepsilon-DP.) For DP, the amount of noise necessary changes with the size nn of the private dataset. For MIP, the amount of noise does not change, so there is only one line.

The results show that for even small datasets (n36n\geq 36) and for η0.01\eta\geq 0.01, direct noise accounting for MIP gives a large advantage over guaranteeing MIP via DP. In practice, such small datasets are uncommon. As nn increases above even this modest range, the advantage in terms of noise reduction for MIP vs. DP quickly becomes many orders of magnitude and is not visible on the plot. (Refer to Proposition 9. The noise required for DP grows exponentially in nn, while it remains constant in nn for MIP.)

Refer to caption
Figure 1: Noise level vs. privacy guarantee for MIP and DP (lower is better). For datasets with at least n=36n=36 points and for almost all values of η\eta, MIP allows us to add much less noise than what would be required by naively applying DP. For n>48n>48, the amount of noise required by DP is so large that it will not appear on the plot.

5.2 Synthetic Data Generation

We conduct an experiment as a representation scenario for private synthetic data generation. We are given a dataset which consists of i.i.d. draws from a private ground truth distribution PP. Our goal is to learn a generative model GG which allows us to (approximately) sample from PP. That is, given a latent random variable zz (we will take z𝒩(0,I)z\sim\mathcal{N}(0,I)), we have G(z)PG(z)\sim P. If GG is itself private and a good approximation for PP, then by the post-processing theorem (Theorem 3), we can use GG to generate synthetic draws from PP, without violating the privacy of any sample in the training data.

For this experiment, we will take GG to be a single-layer linear network with no bias, i.e. G(z)=WzG(z)=Wz for some weight matrix WW. The ground truth distribution P=𝒩(0,Σ)P=\mathcal{N}(0,\Sigma) is a mean-0 normal distribution with (unknown) non-identity covariance Σ\Sigma. In this setting, W=Σ1/2W=\Sigma^{1/2} would exactly reproduce PP.

Let {𝐱i}i=1nd\{\mathbf{x}_{i}\}_{i=1}^{n}\subseteq\mathbb{R}^{d} be the training data. Rather than attempting to learn WΣ1/2W\approx\Sigma^{1/2} directly, we will instead try to learn AΣA\approx\Sigma. We can then set W=(A+A)1/2W=(A+A^{\top})^{1/2}. If AΣA\approx\Sigma, then we will have WΣ1/2W\approx\Sigma^{1/2} and G(z)PG(z)\approx P. We learn AA by minimizing the objective

minAA1ni=1n𝐱i𝐱iF2\min_{A}\>\left\|A-\frac{1}{n}\sum_{i=1}^{n}\mathbf{x}_{i}\mathbf{x}_{i}^{\top}\right\|_{F}^{2}

via gradient descent.

For the DP method, we implement DP-SGD (Abadi et al., 2016) with a full batch. We chose the clipping parameter CC according to the authors’ recommendations, i.e. the median of the unclipped gradients over the training procedure. To implement MIP, we used Corollary 8 with M{2,4,6}M\in\{2,4,6\} and the corresponding σi\sigma_{i}’s computed empirically over 128 random train/holdout splits of the base dataset. The results below use d=3d=3 and n=500,000n=500,000.

Refer to Figure 2. The xx-axes show the theoretical privacy level η\eta (again using the tight correspondence between ε\varepsilon and η\eta from Theorem 4), and the yy-axes show the relative error AΣF/ΣF\|A-\Sigma\|_{F}/\|\Sigma\|_{F} (lower is better). The three plots show the same results zoomed in on different ranges for η\eta to see greater granularity, and the shaded region shows the standard error of the mean over 10 runs. For η0.1\eta\geq 0.1, MIP with any of the tested values of MM outperforms DP in terms of accuracy. For the entire tested η\eta range (the smallest of which was η=0.01\eta=0.01), MIP with M=4M=4 or 66 outperforms DP. (The first plot does not show MIP with M=2M=2 because its error is very large when η<0.1\eta<0.1.) Finally, observe that DP is never able to obtain relative error less than 11 (meaning the resulting output is mostly noise), while MIP obtains relative error less than 11 for η0.2\eta\approx 0.2 and larger.

Refer to caption
Figure 2: Error vs. privacy guarantee for MIP and DP (lower is better). Raw refers to the error of the non-private base algorithm, which just computes AA by vanilla gradient descent. MIP improves over DP in terms of accuracy when η0.1\eta\geq 0.1 for all of the tested values of M{2,4,6}M\in\{2,4,6\}, and for M{4,6}M\in\{4,6\}, MIP improves of DP for the entire range of η\eta. Note that MIP obtains relative error <1<1 for some η\eta, while DP always has relative error larger than 11. MIP with M=2M=2 is not shown in the first plot because the error is large in the small η\eta range, obscuring the results for the other methods.

6 Discussion

In this work, we proposed a novel privacy property, membership inference privacy (MIP) and explained its properties and relationship with differential privacy (DP). The MIP property is more readily interpretable than the guarantees offered by (DP). MIP also requires a smaller amount of noise to guarantee as compared to DP, and therefore can retain greater utility in practice. We proposed a simple “wrapper” method for guaranteeing MIP, which can be implemented with a minor modification both to simple statistical queries or more complicated tasks such as the training procedure for parametric machine learning models.

Limitations

As the example used to prove Theorem 5 shows, there are cases where apparently non-private algorithms can satisfy MIP. Thus, algorithms which satisfy MIP may require post-processing to ensure that the output is not one of the low-probability events in which data privacy is leaked. In addition, because MIP is determined with respect to a holdout set still drawn from 𝒟\mathcal{D}, an adversary may be able to determine with high probability whether or not a given sample was contained in 𝒟\mathcal{D}, rather than just in 𝒟train\mathcal{D}_{\mathrm{train}}, if 𝒟\mathcal{D} is sufficiently different from the rest of the population.

Future Work

Theorem 4 suggests that DP implies MIP in general. However, Theorem 7 shows that a finer-grained analysis of a standard DP mechanism (the Laplace mechanism) is possible, showing that we can guarantee MIP with less noise. It seems plausible that a similar analysis can be undertaken for other DP mechanisms. In addition to these “wrapper” type methods which can be applied on top of existing algorithms, bespoke algorithms for guaranteeing MIP in particular applications (such as synthetic data generation) are also of interest. Noise addition is a simple and effective way to enforce privacy, but other classes of mechanisms may also be possible. For instance, is it possible to directly regularize a probabilistic model using Proposition 2? Finally, the connections between MIP and other theoretical notions of privacy (Renyi DP (Mironov, 2017), concentrated DP (Dwork and Rothblum, 2016), etc.) are also of interest. Lastly, this paper focused on developing on the theoretical principles and guarantees of MIP, but systematic empirical evaluation is an important direction for future work.

References

  • Abadi et al. (2016) Martin Abadi, Andy Chu, Ian Goodfellow, H Brendan McMahan, Ilya Mironov, Kunal Talwar, and Li Zhang. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC conference on computer and communications security, pages 308–318, 2016.
  • Assefa et al. (2020) Samuel A Assefa, Danial Dervovic, Mahmoud Mahfouz, Robert E Tillman, Prashant Reddy, and Manuela Veloso. Generating synthetic data in finance: opportunities, challenges and pitfalls. In Proceedings of the First ACM International Conference on AI in Finance, pages 1–8, 2020.
  • Baluta et al. (2022) Teodora Baluta, Shiqi Shen, S Hitarth, Shruti Tople, and Prateek Saxena. Membership inference attacks and generalization: A causal perspective. ACM SIGSAC Conference on Computer and Communications Security, 2022.
  • Bun et al. (2021) Mark Bun, Damien Desfontaines, Cynthia Dwork, Moni Naor, Kobbi Nissim, Aaron Roth, Adam Smith, Thomas Steinke, Jonathan Ullman, and Salil Vadhan. Statistical inference is not a privacy violation. DifferentialPrivacy.org, 06 2021. https://differentialprivacy.org/inference-is-not-a-privacy-violation/.
  • Carlini et al. (2021a) Nicholas Carlini, Steve Chien, Milad Nasr, Shuang Song, Andreas Terzis, and Florian Tramer. Membership inference attacks from first principles. arXiv preprint arXiv:2112.03570, 2021a.
  • Carlini et al. (2021b) Nicholas Carlini, Florian Tramer, Eric Wallace, Matthew Jagielski, Ariel Herbert-Voss, Katherine Lee, Adam Roberts, Tom Brown, Dawn Song, Ulfar Erlingsson, et al. Extracting training data from large language models. In 30th USENIX Security Symposium (USENIX Security 21), pages 2633–2650, 2021b.
  • Chen et al. (2020) Dingfan Chen, Ning Yu, Yang Zhang, and Mario Fritz. Gan-leaks: A taxonomy of membership inference attacks against generative models. In Proceedings of the 2020 ACM SIGSAC conference on computer and communications security, pages 343–362, 2020.
  • Dwork and Rothblum (2016) Cynthia Dwork and Guy N Rothblum. Concentrated differential privacy. arXiv preprint arXiv:1603.01887, 2016.
  • Dwork et al. (2014) Cynthia Dwork, Aaron Roth, et al. The algorithmic foundations of differential privacy. Foundations and Trends® in Theoretical Computer Science, 9(3–4):211–407, 2014.
  • Gerlovina and Hubbard (2019) Inna Gerlovina and Alan E Hubbard. Computer algebra and algorithms for unbiased moment estimation of arbitrary order. Cogent mathematics & statistics, 6(1):1701917, 2019.
  • Gilbert and McMillan (2018) Anna C Gilbert and Audra McMillan. Property testing for differential privacy. In 2018 56th Annual Allerton Conference on Communication, Control, and Computing (Allerton), pages 249–258. IEEE, 2018.
  • He et al. (2022) Xinlei He, Zheng Li, Weilin Xu, Cory Cornelius, and Yang Zhang. Membership-doctor: Comprehensive assessment of membership inference against machine learning models. arXiv preprint arXiv:2208.10445, 2022.
  • Hu et al. (2022) Pingyi Hu, Zihan Wang, Ruoxi Sun, Hu Wang, and Minhui Xue. M^ 4i: Multi-modal models membership inference. Advances in Neural Information Processing Systems, 2022.
  • Jagielski et al. (2020) Matthew Jagielski, Jonathan Ullman, and Alina Oprea. Auditing differentially private machine learning: How private is private sgd? Advances in Neural Information Processing Systems, 33:22205–22216, 2020.
  • Jayaraman and Evans (2019) Bargav Jayaraman and David Evans. Evaluating differentially private machine learning in practice. In 28th USENIX Security Symposium (USENIX Security 19), pages 1895–1912, 2019.
  • Jayaraman and Evans (2022) Bargav Jayaraman and David Evans. Are attribute inference attacks just imputation? ACM SIGSAC Conference on Computer and Communications Security, 2022.
  • Jordon et al. (2022) James Jordon, Lukasz Szpruch, Florimond Houssiau, Mirko Bottarelli, Giovanni Cherubin, Carsten Maple, Samuel N Cohen, and Adrian Weller. Synthetic data–what, why and how? arXiv preprint arXiv:2205.03257, 2022.
  • Kairouz et al. (2015) Peter Kairouz, Sewoong Oh, and Pramod Viswanath. The composition theorem for differential privacy. In International conference on machine learning, pages 1376–1385. PMLR, 2015.
  • Liu et al. (2022) Yiyong Liu, Zhengyu Zhao, Michael Backes, and Yang Zhang. Membership inference attacks by exploiting loss trajectory. ACM SIGSAC Conference on Computer and Communications Security, 2022.
  • Mironov (2017) Ilya Mironov. Rényi differential privacy. In 2017 IEEE 30th computer security foundations symposium (CSF), pages 263–275. IEEE, 2017.
  • Nasr et al. (2021) Milad Nasr, Shuang Songi, Abhradeep Thakurta, Nicolas Papernot, and Nicholas Carlin. Adversary instantiation: Lower bounds for differentially private machine learning. In 2021 IEEE Symposium on security and privacy (SP), pages 866–882. IEEE, 2021.
  • Office for Civil Rights (2002) HHS Office for Civil Rights. Standards for privacy of individually identifiable health information. final rule. Federal register, 67(157):53181–53273, 2002.
  • Shokri et al. (2017) Reza Shokri, Marco Stronati, Congzheng Song, and Vitaly Shmatikov. Membership inference attacks against machine learning models. In 2017 IEEE symposium on security and privacy (SP), pages 3–18. IEEE, 2017.
  • Stadler et al. (2022) Theresa Stadler, Bristen Oprisanu, and Carmela Troncoso. Synthetic data – anonymisation groundhog day. In 31st USENIX Security Symposium (USENIX Security 22). USENIX Association, 2022.
  • Tao et al. (2021) Yuchao Tao, Ryan McKenna, Michael Hay, Ashwin Machanavajjhala, and Gerome Miklau. Benchmarking differentially private synthetic data generation algorithms. arXiv preprint arXiv:2112.09238, 2021.
  • Thudi et al. (2022) Anvith Thudi, Ilia Shumailov, Franziska Boenisch, and Nicolas Papernot. Bounding membership inference. arXiv preprint arXiv:2202.12232, 2022.
  • Wasserman and Zhou (2010) Larry Wasserman and Shuheng Zhou. A statistical framework for differential privacy. Journal of the American Statistical Association, 105(489):375–389, 2010.

Appendix A Deferred Proofs

For the reader’s convenience, we restate all lemmas, theorems, etc. here. See 2

Proof.

We will show that the membership inference algorithm (𝐱,A)=𝟙{(𝐱𝒟train|𝒜(𝒟train)=A)1/2}\mathcal{I}(\mathbf{x}^{*},A)=\mathds{1}\{\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}})=A)\geq 1/2\} is optimal, then compute the resulting probability of membership inference. We have

((𝐱,𝒜(𝒟train))=y)\displaystyle\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(\mathcal{D}_{\mathrm{train}}))=y^{*}) =𝒟train𝒟(nk)1A𝔸(𝒜(𝒟train)=A)((𝐱,A)=𝟙{𝐱𝒟train})\displaystyle=\sum_{\mathcal{D}_{\mathrm{train}}\subseteq\mathcal{D}}\binom{n}{k}^{-1}\sum_{A\in\mathbb{A}}\mathbb{P}(\mathcal{A}(\mathcal{D}_{\mathrm{train}})=A)\cdot\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},A)=\mathds{1}\{\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\})
=(nk)1A𝔸[D𝔻in(𝒜(D)=A)((𝐱,A)=1)\displaystyle=\binom{n}{k}^{-1}\sum_{A\in\mathbb{A}}\Bigg{[}\sum_{D\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}(\mathcal{A}(D)=A)\cdot\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},A)=1)
+D𝔻out(𝒜(D)=A)(1((𝐱,A)=1))]\displaystyle\hphantom{=\binom{n}{k}^{-1}\sum_{A\in\mathbb{A}}\Bigg{[}}+\sum_{D\in\mathbb{D}^{\mathrm{out}}}\mathbb{P}(\mathcal{A}(D)=A)\cdot(1-\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},A)=1))\Bigg{]}
=(nk)1A𝔸[(D𝔻in(𝒜(D)=A)D𝔻out(𝒜(D)=A))((𝐱,A)=1)\displaystyle=\binom{n}{k}^{-1}\sum_{A\in\mathbb{A}}\Bigg{[}\left(\sum_{D\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}(\mathcal{A}(D)=A)-\sum_{D\in\mathbb{D}^{\mathrm{out}}}\mathbb{P}(\mathcal{A}(D)=A)\right)\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},A)=1)
+D𝔻out(𝒜(D)=A)].\displaystyle\hphantom{=\binom{n}{k}^{-1}\sum_{A\in\mathbb{A}}\Bigg{[}}+\sum_{D\in\mathbb{D}^{\mathrm{out}}}\mathbb{P}(\mathcal{A}(D)=A)\Bigg{]}.

The choice of algorithm \mathcal{I} just specifies the value of ((𝐱,A)=1)\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},A)=1) for each sample 𝐱\mathbf{x}^{*} and each A𝔸A\in\mathbb{A}. We see that the maximum membership inference probability is obtained when

((𝐱,A)=1)=𝟙{D𝔻in(𝒜(D)=A)D𝔻out(𝒜(D)=A)0},\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},A)=1)=\mathds{1}\left\{\sum_{D\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}(\mathcal{A}(D)=A)-\sum_{D\in\mathbb{D}^{\mathrm{out}}}\mathbb{P}(\mathcal{A}(D)=A)\geq 0\right\}, (2)

which implies that

D𝔻in(𝒜(D)=A)D𝔻out(𝒜(D)=A)(nk)1A𝔸max{D𝔻in(𝒜(D)=A),D𝔻out(𝒜(D)=A)}.\sum_{D\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}(\mathcal{A}(D)=A)-\sum_{D\in\mathbb{D}^{\mathrm{out}}}\mathbb{P}(\mathcal{A}(D)=A)\leq\binom{n}{k}^{-1}\sum_{A\in\mathbb{A}}\max\left\{\sum_{D\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}(\mathcal{A}(D)=A),\sum_{D\in\mathbb{D}^{\mathrm{out}}}\mathbb{P}(\mathcal{A}(D)=A)\right\}. (3)

To conclude, observe that

(𝐱D|𝒜(D)=A)=(xD𝒜(D)=A)(𝒜(D)=A)=D𝔻in(nk)1𝒜(𝒜(D)=A)𝒜,D(𝒜(D)=A).\mathbb{P}(\mathbf{x}^{*}\in D\>|\>\mathcal{A}(D)=A)=\frac{\mathbb{P}(x\in D\wedge\mathcal{A}(D)=A)}{\mathbb{P}(\mathcal{A}(D)=A)}=\frac{\sum_{D\in\mathbb{D}^{\mathrm{in}}}\binom{n}{k}^{-1}\mathbb{P}_{\mathcal{A}}(\mathcal{A}(D)=A)}{\mathbb{P}_{\mathcal{A},D}(\mathcal{A}(D)=A)}. (4)

The result follows by rearranging the expression (4) and plugging it into (2) and (3). ∎

In what follows, we will assume without loss of generality that kn/2k\geq n/2. The proofs in the case k<n/2k<n/2 are almost identical and can be obtained by simply swapping 𝔻in𝔻out\mathbb{D}^{\mathrm{in}}\leftrightarrow\mathbb{D}^{\mathrm{out}} and knkk\leftrightarrow n-k.

Lemma 10.

Fix 𝐱𝒟\mathbf{x}^{*}\in\mathcal{D} and let 𝔻in={D𝒟|𝐱D}\mathbb{D}^{\mathrm{in}}=\{D\in\mathcal{D}\>|\>\mathbf{x}^{*}\in D\} and 𝔻out={D𝒟|𝐱D}\mathbb{D}^{\mathrm{out}}=\{D\in\mathcal{D}\>|\>\mathbf{x}^{*}\not\in D\}. If kn/2k\geq n/2 then there is an injective function f:𝔻out𝔻inf:\mathbb{D}^{\mathrm{out}}\rightarrow\mathbb{D}^{\mathrm{in}} such that Df(D)D\sim f(D) for all D𝔻outD\in\mathbb{D}^{\mathrm{out}}.

Proof.

We define a bipartite graph GG on nodes 𝔻in\mathbb{D}^{\mathrm{in}} and 𝔻out\mathbb{D}^{\mathrm{out}}. There is an edge between Din𝔻inD^{\mathrm{in}}\in\mathbb{D}^{\mathrm{in}} and Dout𝔻outD^{\mathrm{out}}\in\mathbb{D}^{\mathrm{out}} if DoutD^{\mathrm{out}} can be obtained from DinD^{\mathrm{in}} by removing 𝐱\mathbf{x}^{*} from DinD^{\mathrm{in}} and replacing it with another element, i.e. if DinDoutD^{\mathrm{in}}\sim D^{\mathrm{out}}. To prove the lemma, it suffices to show that there is a matching on GG which covers DoutD^{\mathrm{out}}. We will show this via Hall’s marriage theorem.

First, observe that GG is a (k,nk)(k,n-k)-biregular graph. Each Din𝔻inD^{\mathrm{in}}\in\mathbb{D}^{\mathrm{in}} has nkn-k neighbors which are obtained from DinD^{\mathrm{in}} by selecting which of the remaining nkn-k elements to replace 𝐱\mathbf{x}^{*} with; each Dout𝔻outD^{\mathrm{out}}\in\mathbb{D}^{\mathrm{out}} has kk neighbors which are obtained by selecting which of the kk elements in DoutD^{\mathrm{out}} to replace with 𝐱\mathbf{x}^{*}.

Let W𝔻outW\subseteq\mathbb{D}^{\mathrm{out}} and let N(W)𝔻inN(W)\subseteq\mathbb{D}^{\mathrm{in}} denote the neighborhood of WW. We have the following:

|N(W)|\displaystyle|N(W)| =DinN(W)DoutW𝟙{DoutDin}DoutW𝟙{DoutDin}\displaystyle=\sum_{D^{\mathrm{in}}\in N(W)}\frac{\sum_{D^{\mathrm{out}}\in W}\mathds{1}\{D^{\mathrm{out}}\sim D^{\mathrm{in}}\}}{\sum_{D^{\mathrm{out}}\in W}\mathds{1}\{D^{\mathrm{out}}\sim D^{\mathrm{in}}\}}
DinN(W)DoutW𝟙{DoutDin}Dout𝔻out𝟙{DoutDin}\displaystyle\geq\sum_{D^{\mathrm{in}}\in N(W)}\frac{\sum_{D^{\mathrm{out}}\in W}\mathds{1}\{D^{\mathrm{out}}\sim D^{\mathrm{in}}\}}{\sum_{D^{\mathrm{out}}\in\mathbb{D}^{\mathrm{out}}}\mathds{1}\{D^{\mathrm{out}}\sim D^{\mathrm{in}}\}}
=1nkDoutWDinN(W)𝟙{DoutDin}\displaystyle=\frac{1}{n-k}\sum_{D^{\mathrm{out}}\in W}\sum_{D^{\mathrm{in}}\in N(W)}\mathds{1}\{D^{\mathrm{out}}\sim D^{\mathrm{in}}\} (5)
=knk|W|.\displaystyle=\frac{k}{n-k}|W|. (6)

Equation (5) holds since each DinD^{\mathrm{in}} has degree nkn-k and by exchanging the order of summation. Similarly, (6) holds since each DoutD^{\mathrm{out}} has degree kk. When kn/2k\geq n/2, we thus have |N(W)||W||N(W)|\geq|W| for every W𝔻outW\subseteq\mathbb{D}^{\mathrm{out}} and the result follows by Hall’s marriage theorem. ∎

See 4

Proof.

Let f:𝔻out𝔻inf:\mathbb{D}^{\mathrm{out}}\rightarrow\mathbb{D}^{\mathrm{in}} denote the injection guaranteed by Lemma 10. We have

(\displaystyle\mathbb{P}( (𝐱,𝒜(D))=y)=1(nk)[Din𝔻in((𝐱,𝒜(Din))=1)+Dout𝔻out((𝐱,𝒜(Dout))=0)]\displaystyle\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D))=y^{*})=\frac{1}{\binom{n}{k}}\left[\sum_{D^{\mathrm{in}}\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D^{\mathrm{in}}))=1)+\sum_{D^{\mathrm{out}}\in\mathbb{D}^{\mathrm{out}}}\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D^{\mathrm{out}}))=0)\right]
1(nk)[Din𝔻in((𝐱,𝒜(Din))=1)+Dout𝔻out(eε((𝐱,𝒜(f(Dout)))=0)+δ)]\displaystyle\leq\frac{1}{\binom{n}{k}}\left[\sum_{D^{\mathrm{in}}\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D^{\mathrm{in}}))=1)+\sum_{D^{\mathrm{out}}\in\mathbb{D}^{\mathrm{out}}}\left(e^{\varepsilon}\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(f(D^{\mathrm{out}})))=0)+\delta\right)\right]
1(nk)[Din𝔻in(((𝐱,𝒜(Din))=1)+eε((𝐱,𝒜(Din))=0))+δ(n1k)]\displaystyle\leq\frac{1}{\binom{n}{k}}\left[\sum_{D^{\mathrm{in}}\in\mathbb{D}^{\mathrm{in}}}\left(\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D^{\mathrm{in}}))=1)+e^{\varepsilon}\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D^{\mathrm{in}}))=0)\right)+\delta\binom{n-1}{k}\right] (7)
1(nk)[eεDin𝔻in(((𝐱,𝒜(Din))=1)+((𝐱,𝒜(Din))=0))+δ(n1k)]\displaystyle\leq\frac{1}{\binom{n}{k}}\left[e^{\varepsilon}\sum_{D^{\mathrm{in}}\in\mathbb{D}^{\mathrm{in}}}\left(\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D^{\mathrm{in}}))=1)+\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D^{\mathrm{in}}))=0)\right)+\delta\binom{n-1}{k}\right]
=1(nk)[eε(n1k1)+δ(n1k)]=eεkn+δnkn.\displaystyle=\frac{1}{\binom{n}{k}}\left[e^{\varepsilon}\binom{n-1}{k-1}+\delta\binom{n-1}{k}\right]=e^{\varepsilon}\frac{k}{n}+\delta\frac{n-k}{n}.

Here inequality (7) critically uses the fact that ff is injective, so at most one term from the sum over 𝔻out\mathbb{D}^{\mathrm{out}} is added to each term in the sum over 𝔻in\mathbb{D}^{\mathrm{in}}. This completes the proof. ∎

See 5

Proof.

Let 𝒜\mathcal{A} be defined as follows. Given a training set DD, 𝒜(D)\mathcal{A}(D) outputs a random subset of DD where each element is included independently and with probability pp. It is obvious that such an algorithm is not (ε,0)(\varepsilon,0)-DP for any ε<\varepsilon<\infty: if 𝐱D\mathbf{x}\in D, then 𝒜(D){{𝐱}}\mathcal{A}(D)\in\{\{\mathbf{x}\}\} with positive probability. But if we replace 𝐱\mathbf{x} with 𝐱𝐱\mathbf{x}^{\prime}\neq\mathbf{x} and call this adjacent dataset DD^{\prime} (so that 𝐱D\mathbf{x}\not\in D^{\prime}, then 𝒜(D){{𝐱}}\mathcal{A}(D^{\prime})\in\{\{\mathbf{x}\}\} with probability 0. Thus 𝒜\mathcal{A} is not differentially private for any p>0p>0.

We now claim that 𝒜\mathcal{A} is (ε,0)(\varepsilon,0)-MIP for any ε>0\varepsilon>0, provided that pp is small enough. To see this, observe the following. For any identification algorithm \mathcal{I},

(\displaystyle\mathbb{P}( (𝐱,𝒜(D))=y)=𝒜(D)[(𝐱D)(𝒜(D)|𝐱D)((𝐱,𝒜(D))=1)\displaystyle\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D))=y^{*})=\sum_{\mathcal{A}(D)}\bigg{[}\mathbb{P}(\mathbf{x}^{*}\in D)\cdot\mathbb{P}(\mathcal{A}(D)\>|\>\mathbf{x}^{*}\in D)\cdot\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D))=1)
+(𝐱D)(𝒜(D)|𝐱D)(1((𝐱,𝒜(D))=1))]\displaystyle\hskip 72.26999pt+\mathbb{P}(\mathbf{x}^{*}\not\in D)\cdot\mathbb{P}(\mathcal{A}(D)\>|\>\mathbf{x}^{*}\not\in D)\cdot(1-\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D))=1))\bigg{]}
=𝒜(D)[(kn(𝒜(D)|𝐱D)+(1kn)(𝒜(D)|𝐱D))((𝐱,𝒜(D))=1)\displaystyle=\sum_{\mathcal{A}(D)}\bigg{[}\left(\frac{k}{n}\mathbb{P}(\mathcal{A}(D)\>|\>\mathbf{x}^{*}\in D)+(1-\frac{k}{n})\mathbb{P}(\mathcal{A}(D)\>|\>\mathbf{x}^{*}\not\in D)\right)\mathbb{P}(\mathcal{I}(\mathbf{x}^{*},\mathcal{A}(D))=1)
+(1kn)(𝒜(D)|𝐱D)]\displaystyle+(1-\frac{k}{n})\mathbb{P}(\mathcal{A}(D)\>|\>\mathbf{x}^{*}\not\in D)\bigg{]}
𝒜(D)max{kn(𝒜(D)|𝐱D),(1kn)(𝒜(D)|𝐱D)}\displaystyle\leq\sum_{\mathcal{A}(D)}\max\left\{\frac{k}{n}\mathbb{P}(\mathcal{A}(D)\>|\>\mathbf{x}^{*}\in D),(1-\frac{k}{n})\mathbb{P}(\mathcal{A}(D)\>|\>\mathbf{x}^{*}\not\in D)\right\}
(1p)kmax{kn,1kn}+𝒜(D)(1(1p)k)max{kn,1kn}\displaystyle\leq(1-p)^{k}\max\left\{\frac{k}{n},1-\frac{k}{n}\right\}+\sum_{\mathcal{A}(D)\neq\emptyset}(1-(1-p)^{k})\max\left\{\frac{k}{n},1-\frac{k}{n}\right\} (8)
=max{kn,1kn}[(1p)k+Cn,k(1(1p)k)].\displaystyle=\max\left\{\frac{k}{n},1-\frac{k}{n}\right\}\left[(1-p)^{k}+C_{n,k}(1-(1-p)^{k})\right]. (9)

Inequality (8) holds because 𝒜(D)=\mathcal{A}(D)=\emptyset with probability (1p)k(1-p)^{k} regardless of whether of not 𝐱D\mathbf{x}^{*}\in D, and therefore the probability that 𝒜(D)\mathcal{A}(D)\neq\emptyset is at most 1(1p)k1-(1-p)^{k} (again regardless of 𝐱D\mathbf{x}^{*}\in D or not) for any 𝒜(D)\mathcal{A}(D)\neq\emptyset. The constant Cn,kC_{n,k} simply counts the number of possible 𝒜(D)\mathcal{A}(D)\neq\emptyset, which depends only on nn and kk but not pp. Thus as p0p\rightarrow 0, (9)1\eqref{eq: no dp 2}\rightarrow 1. This completes the proof. ∎

The proof of Theorem 5 emphasizes that the membership inference privacy guarantee is marginal over the ouput of 𝒜\mathcal{A}. Conditional on a particular output, an adversary may be able to determine whether or not 𝐱D\mathbf{x}^{*}\in D with arbitrarily high precision. This is in contrast with the result of Proposition 6, which shows that even conditionally on a particular output of a DP algorithm, the adversary cannot gain too much.

See 6

Proof.

Using expression (4) (and the corresponding expression for 𝐱𝒟train\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}}), we have

(𝐱𝒟train|𝒜(𝒟train)=A)(𝐱𝒟train|𝒜(𝒟train))\displaystyle\frac{\mathbb{P}(\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}})=A)}{\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\>|\>\mathcal{A}(\mathcal{D}_{\mathrm{train}}))} =D𝔻out𝒜(𝒜(D)=A)D𝔻in𝒜(𝒜(D)=A)\displaystyle=\frac{\sum_{D\in\mathbb{D}^{\mathrm{out}}}\mathbb{P}_{\mathcal{A}}(\mathcal{A}(D)=A)}{\sum_{D\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}_{\mathcal{A}}(\mathcal{A}(D)=A)}
eεD𝔻outminD𝔻in,DD𝒜(𝒜(D)=A)D𝔻in𝒜(𝒜(D)=A).\displaystyle\leq\frac{e^{\varepsilon}\sum_{D\in\mathbb{D}^{\mathrm{out}}}\min_{D^{\prime}\in\mathbb{D}^{\mathrm{in}},D^{\prime}\sim D}\mathbb{P}_{\mathcal{A}}(\mathcal{A}(D)=A)}{\sum_{D\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}_{\mathcal{A}}(\mathcal{A}(D)=A)}.

We now analyze this latter expression. We refer again to the biregular graph GG defined in Lemma 10. For D𝔻outD\in\mathbb{D}^{\mathrm{out}}, N(D)𝔻inN(D)\subseteq\mathbb{D}^{\mathrm{in}} refers to the neighbors of DD in GG, and recall that |N(D)|=k|N(D)|=k for all D𝔻outD\in\mathbb{D}^{\mathrm{out}}. Note that since each D𝔻inD^{\prime}\in\mathbb{D}^{\mathrm{in}} has nkn-k neighbors, we have

D𝔻outDN(D)(𝒜(D)=A)=(nk)D𝔻in(𝒜(D)=A).\sum_{D\in\mathbb{D}^{\mathrm{out}}}\sum_{D^{\prime}\in N(D)}\mathbb{P}(\mathcal{A}(D^{\prime})=A)=(n-k)\sum_{D^{\prime}\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}(\mathcal{A}(D^{\prime})=A).

Using this equality, we have

D𝔻outminD𝔻in,DD𝒜(𝒜(D)=A)D𝔻in𝒜(𝒜(D)=A)\displaystyle\frac{\sum_{D\in\mathbb{D}^{\mathrm{out}}}\min_{D^{\prime}\in\mathbb{D}^{\mathrm{in}},D^{\prime}\sim D}\mathbb{P}_{\mathcal{A}}(\mathcal{A}(D)=A)}{\sum_{D\in\mathbb{D}^{\mathrm{in}}}\mathbb{P}_{\mathcal{A}}(\mathcal{A}(D)=A)} =D𝔻outminDN(D)𝒜(𝒜(D)=A)1nkD𝔻outDN(D)𝒜(𝒜(D)=A)kminDN(D)P𝒜(𝒜(D)=A)\displaystyle=\frac{\sum_{D\in\mathbb{D}^{\mathrm{out}}}\min_{D^{\prime}\in N(D)}\mathbb{P}_{\mathcal{A}}(\mathcal{A}(D)=A)}{\frac{1}{n-k}\sum_{D\in\mathbb{D}^{\mathrm{out}}}\underbrace{\sum_{D^{\prime}\in N(D)}\mathbb{P}_{\mathcal{A}}(\mathcal{A}(D)=A)}_{\geq k\min_{D^{\prime}\in N(D)}P_{\mathcal{A}}(\mathcal{A}(D)=A)}}
nkk.\displaystyle\leq\frac{n-k}{k}.

Since (𝐱𝒟train)=(n1k)/(nk)\mathbb{P}(\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}})=\binom{n-1}{k}/\binom{n}{k} and (𝐱𝒟train)=(n1k1)/(nk)\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}})=\binom{n-1}{k-1}/\binom{n}{k}, we have (𝐱𝒟train)(𝐱𝒟train)=nkk.\frac{\mathbb{P}(\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}})}{\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}})}=\frac{n-k}{k}. This completes the proof. ∎

See 7

Proof.

We will assume that k=n/2k=n/2 is an integer. Let N=|𝔻in|=|𝔻out|N=|\mathbb{D}^{\mathrm{in}}|=|\mathbb{D}^{\mathrm{out}}|, and let 𝔻in={D1,,DN}\mathbb{D}^{\mathrm{in}}=\{D_{1},\ldots,D_{N}\} and 𝔻out={D1,,DN}\mathbb{D}^{\mathrm{out}}=\{D^{\prime}_{1},\ldots,D^{\prime}_{N}\}. Define ai=𝒜(Di)a_{i}=\mathcal{A}(D_{i}) for Di𝔻inD_{i}\in\mathbb{D}^{\mathrm{in}} and bj=𝒜(Dj)b_{j}=\mathcal{A}(D^{\prime}_{j}) for Dj𝔻outD^{\prime}_{j}\in\mathbb{D}^{\mathrm{out}}. Let ZZ be a random variable which is uniformly distributed on {ai}{bj}\{a_{i}\}\cup\{b_{j}\}. We may assume without loss of generality that 𝔼Z=0\mathbb{E}Z=0. In what follows, cc, α\alpha, β\beta, and γ\gamma are constants which we will choose later to optimize our bounds. We also make repeated use of the inequalities 1+xex1+x\leq e^{x} for all xx; 11+x1x\frac{1}{1+x}\geq 1-x for all x0x\geq 0; and ex1+2xe^{x}\leq 1+2x and (1x)(1y)1xy(1-x)(1-y)\geq 1-x-y for 0x,y10\leq x,y\leq 1. Let XX have density proportional to exp(1cσX)\exp(-\frac{1}{c\sigma}\|X\|). The posterior likelihood ratio is given by

f(θ^)=def(𝒟train𝔻in|θ^)(𝒟train𝔻out|θ^)=i=1Nexp(1cσθ^ai)j=1Nexp(1cσθ^bj).f(\hat{\theta})\stackrel{{\scriptstyle\tiny{\mathrm{def}}}}{{=}}\frac{\mathbb{P}(\mathcal{D}_{\mathrm{train}}\in\mathbb{D}^{\mathrm{in}}\>|\>\hat{\theta})}{\mathbb{P}(\mathcal{D}_{\mathrm{train}}\in\mathbb{D}^{\mathrm{out}}\>|\>\hat{\theta})}=\frac{\sum_{i=1}^{N}\exp(-\frac{1}{c\sigma}\|\hat{\theta}-a_{i}\|)}{\sum_{j=1}^{N}\exp(-\frac{1}{c\sigma}\|\hat{\theta}-b_{j}\|)}.

We claim that for all θ^\hat{\theta} with θ^γσclogc\|\hat{\theta}\|\leq\gamma\sigma c\log c, 1η2f(θ^)(1η2)11-\frac{\eta}{2}\leq f(\hat{\theta})\leq(1-\frac{\eta}{2})^{-1}. First, suppose that θ^cασ\|\hat{\theta}\|\leq c^{\alpha}\sigma. Then we have:

f(θ^)\displaystyle f(\hat{\theta}) aicασexp[1cσ(θ^+ai)]N\displaystyle\geq\frac{\sum_{\|a_{i}\|\leq c^{\alpha}\sigma}\exp[-\frac{1}{c\sigma}(\|\hat{\theta}\|+\|a_{i}\|)]}{N}
(12cMα)Ne2cα1N\displaystyle\geq\frac{(1-\frac{2}{c^{M\alpha}})N\cdot e^{-2c^{\alpha-1}}}{N}
14cmin(Mα,1α).\displaystyle\geq 1-4c^{-\min(M\alpha,1-\alpha)}. (10)

Otherwise, θ^cασ\|\hat{\theta}\|\geq c^{\alpha}\sigma. We now have the following chain of inequalities:

f(θ^)\displaystyle f(\hat{\theta}) aicασe1cσ(θ^+ai)bjcασe1cσ(θ^bj)+cασ<bi<θ^e1cσ(θ^bj)+biθ^e1cσ(biθ^)\displaystyle\geq\frac{\sum_{\|a_{i}\|\leq c^{\alpha}\sigma}e^{-\frac{1}{c\sigma}(\|\hat{\theta}\|+\|a_{i}\|)}}{\sum_{\|b_{j}\|\leq c^{\alpha}\sigma}e^{-\frac{1}{c\sigma}(\|\hat{\theta}\|-\|b_{j}\|)}+\sum_{c^{\alpha}\sigma<\|b_{i}\|<\|\hat{\theta}\|}e^{-\frac{1}{c\sigma}(\|\hat{\theta}\|-\|b_{j}\|)}+\sum_{\|b_{i}\|\geq\|\hat{\theta}\|}e^{-\frac{1}{c\sigma}(\|b_{i}\|-\|\hat{\theta}\|)}}
=aicασe1cσaibjcασe1cσbj+cασ<bi<θ^e1cσbj+biθ^e1cσ(2θ^bi)\displaystyle=\frac{\sum_{\|a_{i}\|\leq c^{\alpha}\sigma}e^{-\frac{1}{c\sigma}\|a_{i}\|}}{\sum_{\|b_{j}\|\leq c^{\alpha}\sigma}e^{\frac{1}{c\sigma}\|b_{j}\|}+\sum_{c^{\alpha}\sigma<\|b_{i}\|<\|\hat{\theta}\|}e^{\frac{1}{c\sigma}\|b_{j}\|}+\sum_{\|b_{i}\|\geq\|\hat{\theta}\|}e^{\frac{1}{c\sigma}(2\|\hat{\theta}\|-\|b_{i}\|)}}
N(12cMα)ecα1N(ecα1+2cMαe1cσθ^+2σMθ^Me1cσθ^)\displaystyle\geq\frac{N(1-\frac{2}{c^{M\alpha}})e^{-c^{\alpha-1}}}{N\left(e^{c^{\alpha-1}}+\frac{2}{c^{M\alpha}}e^{\frac{1}{c\sigma}\|\hat{\theta}\|}+\frac{2\sigma^{M}}{\|\hat{\theta}\|^{M}}e^{\frac{1}{c\sigma}\|\hat{\theta}\|}\right)}
(12cMα)ecα1ecα1+2cMαeγlogc+2cMαeγlogc\displaystyle\geq\frac{(1-\frac{2}{c^{M}\alpha})e^{-c^{\alpha-1}}}{e^{c^{\alpha-1}}+\frac{2}{c^{M\alpha}}e^{\gamma\log c}+\frac{2}{c^{M\alpha}}e^{\gamma\log c}}
12cMαcα12cα14cγMα\displaystyle\geq 1-2c^{-M\alpha}-c^{\alpha-1}-2c^{\alpha-1}-4c^{\gamma-M\alpha} (11)
19cmin(1α,Mα2γ).\displaystyle\geq 1-9c^{-\min(1-\alpha,M\alpha-2\gamma)}.

Combining this with (10) shows that f(θ^)19cmin(1α,Mαγ)f(\hat{\theta})\geq 1-9c^{-\min(1-\alpha,M\alpha-\gamma)} for all θ^γσclogc\|\hat{\theta}\|\leq\gamma\sigma c\log c.

Next, we must measure the probability of θ^γσclogc\|\hat{\theta}\|\leq\gamma\sigma c\log c. We can lower bound this probability by first conditioning on the value of 𝒟train\mathcal{D}_{\mathrm{train}}:

(θ^γσclogc)\displaystyle\mathbb{P}(\|\hat{\theta}\|\leq\gamma\sigma c\log c) =1|𝔻|D𝔻(θ^γσclogc|𝒟train=D)\displaystyle=\frac{1}{|\mathbb{D}|}\sum_{D\in\mathbb{D}}\mathbb{P}(\|\hat{\theta}\|\leq\gamma\sigma c\log c\>|\>\mathcal{D}_{\mathrm{train}}=D)
1|𝔻|𝒜(D)cσ(Xγσclogc𝒜(D))\displaystyle\geq\frac{1}{|\mathbb{D}|}\sum_{\|\mathcal{A}(D)\|\leq c\sigma}\mathbb{P}(\|X\|\leq\gamma\sigma c\log c-\|\mathcal{A}(D)\|)
(11cM)(112exp(γσclogccσcσ))\displaystyle\geq\left(1-\frac{1}{c^{M}}\right)\left(1-\frac{1}{2}\exp\left(-\frac{\gamma\sigma c\log c-c\sigma}{c\sigma}\right)\right)
=(11cM)(1e2cγ)\displaystyle=\left(1-\frac{1}{c^{M}}\right)\left(1-\frac{e}{2}c^{-\gamma}\right)
1cMe2cγ.\displaystyle\geq 1-c^{-M}-\frac{e}{2}c^{-\gamma}.

Note that the exact same logic (reversing the roles of the aia_{i}’s and bjb_{j}’s) shows that f(θ^)(19cmin(1α,Mα2γ))1f(\hat{\theta})\leq(1-9c^{-\min(1-\alpha,M\alpha-2\gamma)})^{-1} with probability at least 1cMe2cγ1-c^{-M}-\frac{e}{2}c^{-\gamma} as well.

Finally, we can invoke the result of Proposition 2. Let Δ=9cmin(1α,Mαγ)\Delta=9c^{-\min(1-\alpha,M\alpha-\gamma)} and note that

1Δf(θ^)(1Δ)1max{(𝐱𝒟train|θ^),(𝐱𝒟train|θ^)}12+Δ2.1-\Delta\leq f(\hat{\theta})\leq(1-\Delta)^{-1}\hskip 10.84006pt\Longrightarrow\hskip 10.84006pt\max\left\{\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\>|\>\hat{\theta}),\mathbb{P}(\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}}\>|\>\hat{\theta})\right\}\leq\frac{1}{2}+\frac{\Delta}{2}.

Thus we have

\displaystyle\int max{(𝐱𝒟train|θ^),(𝐱𝒟train|θ^)}d(θ^)\displaystyle\max\left\{\mathbb{P}(\mathbf{x}^{*}\in\mathcal{D}_{\mathrm{train}}\>|\>\hat{\theta}),\mathbb{P}(\mathbf{x}^{*}\not\in\mathcal{D}_{\mathrm{train}}\>|\>\hat{\theta})\right\}\,d\mathbb{P}(\hat{\theta})
(12+Δ2)(f(θ^)[1Δ,(1Δ)1])+(f(θ^)[1Δ,(1Δ)1])\displaystyle\leq\left(\frac{1}{2}+\frac{\Delta}{2}\right)\mathbb{P}(f(\hat{\theta})\in[1-\Delta,(1-\Delta)^{-1}])+\mathbb{P}(f(\hat{\theta})\not\in[1-\Delta,(1-\Delta)^{-1}])
12+92cmin(1α,Mαγ)+2cM+ecγ\displaystyle\leq\frac{1}{2}+\frac{9}{2}c^{-\min(1-\alpha,M\alpha-\gamma)}+2c^{-M}+ec^{-\gamma}
12+(92+e)cmin(1α,Mαγ,γ)+2cM\displaystyle\leq\frac{1}{2}+\left(\frac{9}{2}+e\right)c^{-\min(1-\alpha,M\alpha-\gamma,\gamma)}+2c^{-M} (12)
12+7.5cMM+2\displaystyle\leq\frac{1}{2}+7.5c^{-\frac{M}{M+2}}

where the last inequality follows by setting γ=1α=Mαγ\gamma=1-\alpha=M\alpha-\gamma and solving, yielding γ=M/(M+2)\gamma=M/(M+2). Solving for η=7.5cMM+2\eta=7.5c^{-\frac{M}{M+2}}, we find that c=(7.5η)1+2/Mc=(\frac{7.5}{\eta})^{1+2/M} suffices. This completes the proof. ∎

Corollary 11.

When M2M\geq 2, taking c=(6.16/η)1+2/Mc=(6.16/\eta)^{1+2/M} guarantees η\eta-MIP.

Proof.

The constant improves as MM increases, so it suffices to consider M=2M=2. Let M=2M=2 and α=γ=1/2\alpha=\gamma=1/2, and refer to the proof of Theorem 7. Equation (11) can be improved to

12cMαcα1(e1)cα14cγMα=1(4+e+2c1/2)c1/21-2c^{-M\alpha}-c^{\alpha-1}-(e-1)c^{\alpha-1}-4c^{\gamma-M\alpha}=1-(4+e+2c^{-1/2})c^{-1/2}

using the inequality ex1+(e1)xe^{x}\leq 1+(e-1)x for 0x10\leq x\leq 1 instead of ex1+2xe^{x}\leq 1+2x, which was used to prove Theorem 7. With Δ=(4+e+2c1/2)c1/2\Delta=(4+e+2c^{-1/2})c^{-1/2}, (12) becomes

12+(4+e+2c1/22+e+2c3/2)c1/2.\frac{1}{2}+\left(\frac{4+e+2c^{-1/2}}{2}+e+2c^{-3/2}\right)c^{-1/2}. (13)

Observe that since η1/2\eta\leq 1/2, when we set c=(6.16/η)2c=(6.16/\eta)^{2}, we always have c(26.16)2c\geq(2\cdot 6.16)^{2}, in which case

4+e+2c1/22+e+2c3/26.1597.\frac{4+e+2c^{-1/2}}{2}+e+2c^{-3/2}\leq 6.1597.

Thus, with c=(6.16/η)2c=(6.16/\eta)^{2}, we have

(13)12+6.1597η6.1612+η.\eqref{eq: tight 3}\leq\frac{1}{2}+6.1597\cdot\frac{\eta}{6.16}\leq\frac{1}{2}+\eta.

This completes the proof. ∎

See 8

Proof.

We with to apply the result of Theorem 7 with =σ,M\|\cdot\|=\|\cdot\|_{\sigma,M}. To do this, we must bound the resulting σM\sigma^{M} and show that the density of XX has the correct form. First, observe that

σM=𝔼θ𝔼θM=i=1d𝔼|θi𝔼θi|MdσiMi=1d1d=1.\sigma^{M}=\mathbb{E}\|\theta-\mathbb{E}\theta\|^{M}=\sum_{i=1}^{d}\mathbb{E}\frac{|\theta_{i}-\mathbb{E}\theta_{i}|^{M}}{d\sigma_{i}^{M}}\leq\sum_{i=1}^{d}\frac{1}{d}=1.

It remains to show that the density has the correct form, i.e. depends on XX only through X\|X\|. This will be the case if the marginal density of UU is uniform. Let p(U)p(U) be the density of UU. Observe that, for any u=us,M=1\|u\|=\|u\|_{s,M}=1, we have that YuY\mapsto u iff Y=suY=su for some s>0s>0. Thus

p(u)\displaystyle p(u) s=0e1σ12(su1/σ1)M++(sud/σd)M𝑑s\displaystyle\propto\int_{s=0}^{\infty}e^{\frac{1}{\sigma_{1}^{2}}(su_{1}/\sigma_{1})^{M}+\cdots+(su_{d}/\sigma_{d})^{M}}\,ds
=s=0esMdu2𝑑s\displaystyle=\int_{s=0}^{\infty}e^{-s^{M}d\|u\|^{2}}\,ds
=s=0esMd𝑑s.\displaystyle=\int_{s=0}^{\infty}e^{-s^{M}d}\,ds.

The last inequality holds because u=1\|u\|=1 is constant. Thus, the density is independent of uu and we can directly apply Theorem 7. ∎

Lemma 12 (Chebyshev’s Inequality).

Let \|\cdot\| be any norm and XX be a random vector with 𝔼X𝔼X2σk\mathbb{E}\|X-\mathbb{E}X\|^{2}\leq\sigma^{k}. Then for any t>0t>0, we have

(X𝔼X>tσ)1/tk.\mathbb{P}(\|X-\mathbb{E}X\|>t\sigma)\leq 1/t^{k}.
Proof.

This follows almost directly from Markov’s inequality:

(X𝔼X>tσ)=(X𝔼Xk>tkσk)𝔼X𝔼Xktkσk1/tk.\mathbb{P}(\|X-\mathbb{E}X\|>t\sigma)=\mathbb{P}(\|X-\mathbb{E}X\|^{k}>t^{k}\sigma^{k})\leq\frac{\mathbb{E}\|X-\mathbb{E}X\|^{k}}{t^{k}\sigma^{k}}\leq 1/t^{k}.