This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Differentially Private Mechanisms for Count Queries111This is a working paper. Comments to improve this work are welcome.

Parastoo Sadeghi, Shahab Asoodeh and Flavio du Pin Calmon
RSEEME, Australian National University, Australia
School of Engineering and Applied Sciences, Harvard University, USA
Emails: parastoo.sadeghi@anu.edu.au, {shahab, flavio}@seas.harvard.edu
Abstract

In this paper, we consider the problem of responding to a count query (or any other integer-valued queries) evaluated on a dataset containing sensitive attributes. To protect the privacy of individuals in the dataset, a standard practice is to add continuous noise to the true count. We design a differentially-private mechanism which adds integer-valued noise allowing the released output to remain integer. As a trade-off between utility and privacy, we derive privacy parameters ε{\varepsilon} and δ\delta in terms of the the probability of releasing an erroneous count under the assumption that the true count is no smaller than half the support size of the noise. We then numerically demonstrate that our mechanism provides higher privacy guarantee compared to the discrete Gaussian mechanism that is recently proposed in the literature.

I Introduction

Conducting nationwide censuses is one of the most important tasks of national or federal agencies around the world. For example, the 2020 Census in the US is currently under way, which is conducted by the US Bureau of the Census every ten years [1]. In Australia, census is conducted by the Australian Bureau of Statistics (ABS) every five years and the next round is due in 2021 [2]. Responding to count queries about census outcomes is one of the most important tasks of such agencies. This can be important for distributing national funds, determining optimal voting districts and conducting various scientific or economic studies of the population. However, this might lead to dire privacy compromises.

In order to protect the privacy of individuals in the census records, the true census results are heavily guarded. Instead, noisy versions may be released to the public in a one-off general-purpose publication or privately disclosed to a data analyst in response to specific queries. There is clearly a tension between releasing an accurate and reliable version of the census outcome and protecting the privacy of individuals in a population. This tension gets aggravated particularly for underrespresented groups in small towns and counties. Oftentimes, their livelihoods and future funding depends on them being accurately accounted for in published census results. At the same, due to being very small in size, they are at the greatest risk of being re-identified even if perturbed versions of their data is published. For an interesting coverage of this issue in the 2010 and 2020 US Census rounds, the reader is referred to a recent article published in the New York Times [3].

Motivated by these applications, in this paper we are interested in formulating and studying optimal tradeoffs between utility and privacy when a data analyst sends a query to a data curator about the number of individuals with a common sensitive attribute in a large dataset. Assuming nn is the true count, the data curator releases a random variable Y=n+ZY=n+Z, where ZZ is an integer-valued noise variable with a suitably-chosen probability distribution.

It is not surprising that the celebrated differential privacy (DP) framework [4] has been applied to integer-valued mechanisms for count queries. In fact, DP is a strong candidate for publishing private 2020 US Census outcomes [5]. Other countries such as Australia and New Zealand are closely watching this space and have already taken steps in evaluating the performance of their existing mechanisms “from the lens of differential privacy” [6].

A standard approach to preserving ε{\varepsilon}-DP is to perturb the count query by adding random noise ZZ with Laplacian distribution whose variance is proportional to a certain property of the query (i.e., sensitivity). Adding continuous noise to an integer-valued query makes the output less interpretable. Geometric distribution, as the discrete counterpart of Laplace distribution, was shown in [7] to be ”optimal” for integer-valued queries with unit sensitivity. The optimality was defined in terms of minimizing the expected cost function (chosen from a family of utility functions with some mild regularity conditions) under a Bayesian framework. Their result, in particular, implies that adding (truncated or folded) geometric noise to the above-mentioned counting query preserves DP while minimizing the probability of error, Pr(Z0)\Pr(Z\neq 0). This setting was extended to the mini-max setting in [8] and to queries with general sensitivity and under the worst case setting in [9]. The design and implementation of geometric distribution on finite-precision machines were investigated in [10].

More recently, the problem of minimizing the LpL_{p}-norm of error of the query output, subject to ε{\varepsilon}-DP, was revisited in [11]. It was argued that despite minimizing the probability of error, the geometric mechanism suffers from a number of limitations. Most importantly, the error p(Yn)p(Y\neq n) depends on the true count nn, leading to uneven accuracy for different queries. In the very high-privacy regime when ε0{\varepsilon}\to 0, this unevenness in the distribution results in overwhelmingly reporting two extreme values Y=0Y=0 or Y=maxnY=\max~{}n (zero count or maximum possible count in the database). To address this issue, they proposed an explicit fair mechanism (EM). However, the noise in the EM method adds bias, that is, 𝔼[Y|n]n\mathbb{E}[Y|n]\neq n. Moreover, EM (and also truncated geometric) mechanism does not restrict the support size of the noise. As such, the output can be any number between the two extremes y=0y=0 and y=maxny=\max n with non-zero probability.

In a different, and quite recent, line of work, a discrete version of the Gaussian density was considered for designing privacy-preserving mechanism for count queries. In particular, it was shown in [12] that adding “discrete” Gaussian noise provides similar privacy and utility guarantees to those obtained by the well-understood Gaussian mechanism. Discrete Gaussian noise is supported over \mathbb{Z} and thus the output of such mechanism can be all integers regardless of the domain of the query. This might lead to unpleasant outcomes in some practical scenarios. For example, according to the New York Times article [3], the population of a county in the 2010 US Census was over-reported by a factor of almost 8 (the true count was around 90, but was reported to be around 920). While post-processing (truncating or folding the probability mass function) can be used to limit the range of discrete Gaussian mechanisms, the question remains whether one can explicitly incorporate a finite support of the noise into the mechanism design process.

In this paper, we adopt a different approach than [11, 12]. We begin by enforcing three basic desired constraints on the noise distribution. These properties include zero bias, fixed hard support, and fixed error probability irrespective of the true count nn. By varying the probability of error we can trade off accuracy with privacy. The hard support for ZZ ensures that for any given nn, the output n+Zn+Z takes value in a pre-specified range with probability one; in particular YY is required to be always non-negative. We first design a noise distribution satisfying above constraints and then use it to construct a differentially-private mechanism (under a mild condition on the dataset). The constraint on the noise support has two implications. First, the values of the noise added must depend on the true count nn (to ensure the non-negativity of the output). Second, the mechanism provides a slightly relaxed privacy guarantee: it is ε{\varepsilon}-DP with probability 1δ1-\delta for some parameter δ>0\delta>0. This framework is known as approximate DP and usually denoted by (ε,δ)({\varepsilon},\delta)-DP [13, 14]. While the proposed noise has similar distribution as discrete Gaussian (with slightly heavier tail), our numerical findings indicate that the privacy performance of our mechanism (in terms of ε{\varepsilon} and δ\delta) is tighter than what would be obtained via discrete Gaussian mechanism with the same variance.

We can summarize our observations about the interplay of privacy and utility parameters as follows. As expected, a lower privacy requirement, i.e., higher ε{\varepsilon}, results in lower δ\delta when the noise support size and probability of releasing an erroneous value are fixed. By moderately increasing the noise support size while keeping ε{\varepsilon} and error probability fixed, we can also significantly improve δ\delta. For example, increasing the noise support from 2D+1=92D+1=9 to 2D+1=132D^{\prime}+1=13 we obtain significantly smaller δ\delta. It turns out that when ε{\varepsilon} and the noise support size are fixed, reducing the error probability (and hence improving data reliability) can be tolerated up a certain threshold without significantly affecting δ\delta. However, further reducing the error probability below such a threshold can have severe effects on the privacy performance through increasing δ\delta beyond typically acceptable levels.

Notation: For an integer NN\in\mathbb{N}, we use [N][N] to denote the set {1,2,,N}\{1,2,\cdots,N\}. For a,ba,b\in\mathbb{Z} and aba\leq b, we use [a:b][a:b] to denote the set {a,a+1,,b}\{a,a+1,\cdots,b\}.

II Problem Formulation

Suppose a dataset of size at most NN is given. A query on the number of entries in the database that possess a certain sensitive attribute is received by the data curator. Assume n[N]n\in[N] is the actual (positive) count of such attribute. To protect the privacy of individuals in the dataset, the data curator releases a noisy version of nn as

Y=n+Z,\displaystyle Y=n+Z, (1)

where ZZ is an integer-valued noise variable. We assume the noise distribution pZ(|n)p_{Z}(\cdot|n) satisfies the following intuitive properties:

  1. P1.

    Fix DD\in\mathbb{N}. Noise variable ZZ takes values in [A:D][-A:D], where Amin{n,D}A\doteq\min\{n,D\}, i.e., pZ(z|n)=0p_{Z}(z|n)=0 if z[A:D]z\notin[-A:D]. This properties ensures that the mechanism’s output is always non-negative. A smaller DD implies a smaller noise support set, which in turn implies a larger accuracy in reporting the query.222Note that the support of the output for a given nn is limited to [nA:n+D][n-A:n+D]. For max{1,ND+1}nN\max\{1,N-D+1\}\leq n\leq N, we allow the output to be greater than the maximum possible real count NN (by at most DD). If NDN\gg D, this overshoot is negligible.

  2. P2.

    Fix η(0,1)\eta\in(0,1). The probability of releasing the correct value is set to

    pY(Y=n)=η,p_{Y}(Y=n)=\eta,

    irrespective of the true count nn. Equivalently, the probability of releasing an erroneous value is

    pY(Yn)=1ηη¯.p_{Y}(Y\neq n)=1-\eta\doteq\bar{\eta}.

    A larger η\eta means more reliability in releasing the correct value. Conversely, a smaller η\eta means better privacy in the sense of concealing the true count of the sensitive attribute in the dataset.

  3. P3.

    In addition, the error must have zero bias. That is, 𝔼[Z|n]=0\mathbb{E}[Z|n]=0, which is a statistically desirable property.333Note that if n=0n=0, maintaining zero bias is not possible jointly with η<1\eta<1 and P1. Hence, we limit ourselves to queries with n1n\geq 1.

The following proposition provides a simple family of noise distribution that satisfy these properties.

Proposition 1.

Let the true count for a query be n[N]n\in[N]. For any i1[A:1]i_{1}\in[-A:-1] and i2[D]i_{2}\in[D], the following distribution satisfies P1-P3

pZ(i1,i2)(z|n)={η¯i2i2+|i1|,z=i1,η,z=0,η¯|i1|i2+|i1|,z=i2,0,otherwise.\displaystyle p^{(i_{1},i_{2})}_{Z}(z|n)=\begin{cases}\bar{\eta}\frac{i_{2}}{i_{2}+|i_{1}|},\quad&z=i_{1},\\ \eta,\quad&z=0,\\ \bar{\eta}\frac{|i_{1}|}{i_{2}+|i_{1}|},\quad&z=i_{2},\\ 0,\quad&\text{otherwise}.\end{cases} (2)

In light of this proposition, we can obtain that any convex combination of pZ(i1,i2)(z|n)p^{(i_{1},i_{2})}_{Z}(z|n), i.e.,

pZ(z|n)i1=A1i2=1Dαi1,i2,npZ(i1,i2)(z|n),\displaystyle p_{Z}(z|n)\doteq\sum_{i_{1}=-A}^{-1}\sum_{i_{2}=1}^{D}\alpha_{i_{1},i_{2},n}\,\,p^{(i_{1},i_{2})}_{Z}(z|n), (3)

with non-negative coefficients αi1,i2,n\alpha_{i_{1},i_{2},n} satisfying i1=A1i2=1Dαi1,i2,n=1\sum_{i_{1}=-A}^{-1}\sum_{i_{2}=1}^{D}\alpha_{i_{1},i_{2},n}=1, also meets P1-P3. We can thus use this convex combination to obtain a general data-dependent noise distribution satisfying P1-P3. The following example makes this construction clear.

Example 1.

Let N=3N=3, η=12\eta=\frac{1}{2} and D=2D=2. Note that for n=1n=1, two distributions are possible for pZ(i1,i2)(z|1)p^{(i_{1},i_{2})}_{Z}(z|1) by choosing (i1,i2)=(1,1)(i_{1},i_{2})=(-1,1) and (i1,i2)=(1,2)(i_{1},i_{2})=(-1,2), as described below

pZ(1,1)(z|1)={14,z=1,112,z=0,0,otherwise,andpZ(1,2)(z|1)={13,z=1,12,z=0,16,z=2,0,otherwise.p^{(-1,1)}_{Z}(z|1)=\begin{cases}\frac{1}{4},\quad&z=-1,1\\ \frac{1}{2},\quad&z=0,\\ 0,\quad&\text{otherwise},\end{cases}\qquad\text{and}\qquad p^{(-1,2)}_{Z}(z|1)=\begin{cases}\frac{1}{3},\quad&z=-1,\\ \frac{1}{2},\quad&z=0,\\ \frac{1}{6},\quad&z=2,\\ 0,\quad&\text{otherwise}.\end{cases} (4)

For n=2n=2 and n=3n=3, four distributions are similarly possible by combining i1=2,1i_{1}=-2,-1 with i2=1,2i_{2}=1,2. Overall, the general convex structure in (3) produces the noise distributions pZ(|1),pZ(|2),p_{Z}(\cdot|1),p_{Z}(\cdot|2), and pZ(|3)p_{Z}(\cdot|3) given by

pZ(z|n)=(n=1n=2n=3014α2,2,2+16α2,1,214α2,2,3+16α2,1,314α1,1,1+13α1,2,114α1,1,2+13α1,2,214α1,1,3+13α1,2,312121214α1,1,114α1,1,2+13α2,1,214α1,1,3+13α2,1,316α1,2,114α2,2,2+16α1,2,214α2,2,3+16α1,2,3).\displaystyle p_{Z}(z|n)=\left(\begin{matrix}n=1&&n=2&&n=3\\ \hline\cr\\ 0&&\frac{1}{4}\alpha_{-2,2,2}+\frac{1}{6}\alpha_{-2,1,2}&&\frac{1}{4}\alpha_{-2,2,3}+\frac{1}{6}\alpha_{-2,1,3}\\ \frac{1}{4}\alpha_{-1,1,1}+\frac{1}{3}\alpha_{-1,2,1}&&\frac{1}{4}\alpha_{-1,1,2}+\frac{1}{3}\alpha_{-1,2,2}&&\frac{1}{4}\alpha_{-1,1,3}+\frac{1}{3}\alpha_{-1,2,3}\\ \frac{1}{2}&&\frac{1}{2}&&\frac{1}{2}\\ \frac{1}{4}\alpha_{-1,1,1}&&\frac{1}{4}\alpha_{-1,1,2}+\frac{1}{3}\alpha_{-2,1,2}&&\frac{1}{4}\alpha_{-1,1,3}+\frac{1}{3}\alpha_{-2,1,3}\\ \frac{1}{6}\alpha_{-1,2,1}&&\frac{1}{4}\alpha_{-2,2,2}+\frac{1}{6}\alpha_{-1,2,2}&&\frac{1}{4}\alpha_{-2,2,3}+\frac{1}{6}\alpha_{-1,2,3}\end{matrix}\right). (5)

Each column of this matrix is a valid noise distribution for a given n[N]n\in[N], subject to i1=A1i2=1Dαi1,i2,n=1\sum_{i_{1}=-A}^{-1}\sum_{i_{2}=1}^{D}\alpha_{i_{1},i_{2},n}=1. Note that row indices range from z=Dz=-D to z=Dz=D. Also, note that pZ(2|1)=0p_{Z}(-2|1)=0. This is because for n=1n=1, we have A=min{n,D}=1A=\min\{n,D\}=1, dictating [1:2][-1:2] as the support set of p1p_{1}, instead of [2:2][-2:2]. Moreover, pZ(z|n)=0p_{Z}(z|n)=0 for n4n\geq 4 or z[2:2]z\notin[-2:2]. By adding such noise to the true count nn, the distribution of YY, given nn, becomes

pY(y|n)=(n=1n=2n=314α1,1,1+13α1,2,114α2,2,2+16α2,1,201214α1,1,2+13α1,2,214α2,2,3+16α2,1,314α1,1,11214α1,1,3+13α1,2,316α1,2,114α1,1,2+13α2,1,212014α2,2,2+16α1,2,214α1,1,3+13α2,1,30014α2,2,3+16α1,2,3),\displaystyle p_{Y}(y|n)=\left(\begin{matrix}n=1&&n=2&&n=3\\ \hline\cr\\ \frac{1}{4}\alpha_{-1,1,1}+\frac{1}{3}\alpha_{-1,2,1}&&\frac{1}{4}\alpha_{-2,2,2}+\frac{1}{6}\alpha_{-2,1,2}&&0\\ \frac{1}{2}&&\frac{1}{4}\alpha_{-1,1,2}+\frac{1}{3}\alpha_{-1,2,2}&&\frac{1}{4}\alpha_{-2,2,3}+\frac{1}{6}\alpha_{-2,1,3}\\ \frac{1}{4}\alpha_{-1,1,1}&&\frac{1}{2}&&\frac{1}{4}\alpha_{-1,1,3}+\frac{1}{3}\alpha_{-1,2,3}\\ \frac{1}{6}\alpha_{-1,2,1}&&\frac{1}{4}\alpha_{-1,1,2}+\frac{1}{3}\alpha_{-2,1,2}&&\frac{1}{2}\\ 0&&\frac{1}{4}\alpha_{-2,2,2}+\frac{1}{6}\alpha_{-1,2,2}&&\frac{1}{4}\alpha_{-1,1,3}+\frac{1}{3}\alpha_{-2,1,3}\\ 0&&0&&\frac{1}{4}\alpha_{-2,2,3}+\frac{1}{6}\alpha_{-1,2,3}\end{matrix}\right), (6)

where the rows range from y=0y=0 to y=N+D=5y=N+D=5. The value of pY(y|n)p_{Y}(y|n) for all other values of yy and nn not shown above is identical to zero.

II-A Incorporating (ε,δ)({\varepsilon},\delta)-DP Conditions

Consider the mechanism Y=n+ZY=n+Z with ZZ being a data-dependant noise with distribution pZ(|n)p_{Z}(\cdot|n) given in (3). Our objective is to determine the smallest ε0{\varepsilon}\geq 0 and δ(0,1)\delta\in(0,1) such that this mechanism is (ε,δ)({\varepsilon},\delta)-differentially private (DP) [13].

Definition 1.

Given ε0{\varepsilon}\geq 0 and δ[0,1)\delta\in[0,1), a randomized algorithm M:𝒳N𝒴M:\mathcal{X}^{N}\to\mathcal{Y} is said to be (ε,δ)({\varepsilon},\delta)-differentially private (DP) if p(M(x)S)eεp(M(x)S)+δp(M(x)\in S)\leq e^{\varepsilon}p(M(x^{\prime})\in S)+\delta for all S𝒴S\subset\mathcal{Y} and all neighboring datasets x,x𝒳Nx,x^{\prime}\in\mathcal{X}^{N} differing in one element.

Below we specialize this standard definition for our count query setup. Given an integer-valued query q:𝒳N[N]q:\mathcal{X}^{N}\to[N] about a dataset xx, mechanism qq randomizes the true query response, say nn, via a data-dependent additive noise process, i.e., q(x)=n+Zq(x)=n+Z. Since the query is integer-valued, we have |q(x)q(x)|1|q(x)-q(x^{\prime})|\leq 1 for all pairs of neighboring datasets xx and xx^{\prime}. Without loss of generality, in the following we assume that |q(x)q(x)|=1|q(x)-q(x^{\prime})|=1 for neighboring xx and xx in the computation of DP parameters. Assuming q(x)=nq(x)=n, then it implies that we need to verify the DP requirement for dataset xx^{\prime} for which q(x)=n+1q(x^{\prime})=n+1 or q(x)=n1q(x^{\prime})=n-1.

Remark 1.

Our approach to computing the DP parameters of mechanism (16) is as follows. We first consider Definition 1 for the singular events S[0:N+D]S\subset[0:N+D], i.e., all subsets SS satisfying |S|=1|S|=1 and determine the corresponding parameters ε{\varepsilon} and δ\delta. Clearly, these parameters differ from the DP parameters. We then convert these parameters to a valid set of DP parameters. In the following, we use ε{\varepsilon} and δ\delta to denote the parameters of mechanism (16) when applying Definition 1 for singular events SS.

To formulate our goal, we fix ε0{\varepsilon}\geq 0 and consider the following linear program

δ\displaystyle\delta^{*}\doteq minδ,\displaystyle~{}\min~{}\delta,
s.t.pY(y|n)eεpY(y|n+1)+δ,n[1:N1],y[0:N+D],\displaystyle~{}~{}\text{s.t.}~{}p_{Y}(y|n)\leq e^{\varepsilon}p_{Y}(y|n+1)+\delta,~{}~{}\forall n\in[1:N-1],~{}~{}\forall y\in[0:N+D], (7)
pY(y|n)eεpY(y|n1)+δ,n[2:N],y[0:N+D].\displaystyle~{}~{}~{}~{}~{}~{}p_{Y}(y|n)\leq e^{\varepsilon}p_{Y}(y|n-1)+\delta,~{}~{}\forall n\in[2:N],~{}~{}\forall y\in[0:N+D]. (8)

Given the mechanism Y=n+ZY=n+Z and Proposition 1, we can explicitly write the optimization of δ\delta in terms of αi1,i2,n\alpha_{i_{1},i_{2},n} as follows.

Proposition 2.

For given ε0{\varepsilon}\geq 0, the optimal value δ\delta^{*} is the solution of the following linear program

minδ\displaystyle\min\delta (9)
s.t.
i1,i2αi1,i2,npZ(i1,i2)(yn|n)\displaystyle\sum_{i_{1},i_{2}}\alpha_{i_{1},i_{2},n}\,\,p^{(i_{1},i_{2})}_{Z}(y-n|n) eεi1,i2αi1,i2,n+1pZ(i1,i2)(yn1|n+1)+δ,n[1:N1],\displaystyle\leq e^{\varepsilon}\sum_{i_{1},i_{2}}\alpha_{i_{1},i_{2},n+1}\,\,p^{(i_{1},i_{2})}_{Z}(y-n-1|n+1)+\delta,~{}\forall n\in[1:N-1], (10)
i1,i2αi1,i2,npZ(i1,i2)(yn|n)\displaystyle\sum_{i_{1},i_{2}}\alpha_{i_{1},i_{2},n}\,\,p^{(i_{1},i_{2})}_{Z}(y-n|n) eεi1,i2αi1,i2,n1pZ(i1,i2)(yn+1|n1)+δ,n[2:N],\displaystyle\leq e^{\varepsilon}\sum_{i_{1},i_{2}}\alpha_{i_{1},i_{2},n-1}\,\,p^{(i_{1},i_{2})}_{Z}(y-n+1|n-1)+\delta,~{}\forall n\in[2:N], (11)
αi1,i2,n\displaystyle\alpha_{i_{1},i_{2},n} 0,i1[A:1],i2[D],n[1:N],\displaystyle\geq 0,\qquad\forall i_{1}\in[-A:-1],~{}\forall i_{2}\in[D],~{}\forall n\in[1:N], (12)
i1=A1i2=1Dαi1,i2,n\displaystyle\sum_{i_{1}=-A}^{-1}\sum_{i_{2}=1}^{D}\alpha_{i_{1},i_{2},n} =1,n[1:N],\displaystyle=1,\qquad\forall n\in[1:N], (13)

where pZ(i1,i2)(|n)p^{(i_{1},i_{2})}_{Z}(\cdot|n) was defined in (2) and y[0:N+D]y\in[0:N+D].

Clearly, this above optimization problem is a linear program. If the computational complexity allows, an explicit expression for δ\delta^{*} can be obtained through Fourier-Motzkin elimination of all αi1,i2,n\alpha_{i_{1},i_{2},n}. Nevertheless, we show in the next section that such expression can be derived directly if we make a minor assumption on DD.

III Explicit Solutions for nDn\geq D

In this section, we show that if the true count satisfies nDn\geq D, then δ\delta^{*} can be readily derived.

Proposition 3.

If noise support parameter DD satisfies DnD\leq n, then the following data-independent noise distribution satisfies P1-P3

pZ(z)={αiη¯2,z=i,iη,z=0,0,otherwise,\displaystyle p_{Z}(z)=\begin{cases}\alpha_{i}\frac{\bar{\eta}}{2},\quad&z=-i,i\\ \eta,\quad&z=0,\\ 0,\quad&\text{otherwise},\end{cases} (14)

for all i[1:D]i\in[1:D] and αi0\alpha_{i}\geq 0 such that i=1Dαi=1\sum_{i=1}^{D}\alpha_{i}=1. In this case,

pY(y|n)={αiη¯2,y=ni,n+iη,y=n,0,otherwise.\displaystyle p_{Y}(y|n)=\begin{cases}\alpha_{i}\frac{\bar{\eta}}{2},\quad&y=n-i,n+i\\ \eta,\quad&y=n,\\ 0,\quad&\text{otherwise}.\end{cases} (15)

Notice that the assumption nDn\geq D enables us to design a noise distribution that satisfies P1-P3 while being independent of the query (i.e., nn). The simple mechanism given in Proposition 3 can be represented by the matrix

pY(y|n)=(n=Dn=D+1n=NαDη¯200αD1η¯2αDη¯20α1η¯2α2η¯20ηα1η¯2αDη¯2α1η¯2ηαD1η¯2αDη¯2αD1η¯2η0αDη¯2α1η¯200000αDη¯2),\displaystyle p_{Y}(y|n)=\left(\begin{matrix}n=D&&n=D+1&&\cdots&&n=N\\ \hline\cr\\ \alpha_{D}\frac{\bar{\eta}}{2}&&0&&\cdots&&0\\ \alpha_{D-1}\frac{\bar{\eta}}{2}&&\alpha_{D}\frac{\bar{\eta}}{2}&&\cdots&&0\\ \cdots&&\cdots&&\cdots&&\cdots\\ \alpha_{1}\frac{\bar{\eta}}{2}&&\alpha_{2}\frac{\bar{\eta}}{2}&&\cdots&&0\\ \eta&&\alpha_{1}\frac{\bar{\eta}}{2}&&\cdots&&\alpha_{D}\frac{\bar{\eta}}{2}\\ \alpha_{1}\frac{\bar{\eta}}{2}&&\eta&&\cdots&&\alpha_{D-1}\frac{\bar{\eta}}{2}\\ \cdots&&\cdots&&\cdots&&\cdots\\ \alpha_{D}\frac{\bar{\eta}}{2}&&\alpha_{D-1}\frac{\bar{\eta}}{2}&&\cdots&&\eta\\ 0&&\alpha_{D}\frac{\bar{\eta}}{2}&&\cdots&&\alpha_{1}\frac{\bar{\eta}}{2}\\ 0&&0&&\cdots&&\cdots\\ 0&&0&&0&&\alpha_{D}\frac{\bar{\eta}}{2}\end{matrix}\right), (16)

where for exposition we have assumed N=2DN=2D. Each column of this matrix represents pY(y|n)p_{Y}(y|n) for y[0:3D]y\in[0:3D].

III-A Lower bounds on δ\delta^{*}

To compute δ\delta^{*} for mechanism (16), we first derive a lower bound for δ\delta^{*} (this subsection) and then show that it is in fact achievable (next subsection). For notational brevity, define EeεE\doteq e^{\varepsilon}. The main tool for deriving the lower bound is the following proposition.

Proposition 4.

For the noise distribution given by Proposition 3, δ\delta^{*} satisfies the following inequalities

η¯2αD\displaystyle\frac{\bar{\eta}}{2}\alpha_{D} δ,\displaystyle\leq\delta^{*}, (17)
η¯2αi\displaystyle\frac{\bar{\eta}}{2}\alpha_{i} Eη¯2αi+1+δ,i[1:D1],\displaystyle\leq E\frac{\bar{\eta}}{2}\alpha_{i+1}+\delta^{*},\quad i\in[1:D-1], (18)
η\displaystyle\eta Eη¯2α1+δ.\displaystyle\leq E\frac{\bar{\eta}}{2}\alpha_{1}+\delta^{*}. (19)

This proposition follows simply by comparing consecutive columns of the matrix (16). Letting B2η¯B\doteq\frac{2}{\bar{\eta}} and C2ηη¯C\doteq\frac{2\eta}{\bar{\eta}}, the system of inequalities given in Proposition 4 can be expressed as

αD\displaystyle\alpha_{D} Bδ,\displaystyle\leq B\delta^{*}, (20)
αi+1\displaystyle\alpha_{i+1} αiBδE,1iD1,\displaystyle\geq\frac{\alpha_{i}-B\delta^{*}}{E},\qquad 1\leq i\leq D-1, (21)
α1\displaystyle\alpha_{1} CBδE.\displaystyle\geq\frac{C-B\delta^{*}}{E}. (22)

We now present a series of lower bounds on δ\delta^{*} that we term Type-I lower bounds.

Lemma 1.

For the noise distribution given by Proposition 3, we have

δδkCj=0k1EjEkBj=0k1Ej(j+1),k[D].\displaystyle\delta^{*}\geq\delta_{k}\doteq\frac{C\sum_{j=0}^{k-1}E^{j}-E^{k}}{B\sum_{j=0}^{k-1}E^{j}(j+1)},\qquad k\in[D]. (23)
Proof.

First note that, since α11\alpha_{1}\leq 1, we have from (22),

δδ1.\displaystyle\delta^{*}\geq\delta_{1}. (24)

Since α21α1\alpha_{2}\leq 1-\alpha_{1}, a second lower bound is obtained from (21) for i=2i=2 by eliminating both α2\alpha_{2} and α1\alpha_{1} as follows

α1BδEα21α1\displaystyle\frac{\alpha_{1}-B\delta^{*}}{E}\leq\alpha_{2}\leq 1-\alpha_{1}\quad α1(1+E)EBδ\displaystyle\Rightarrow\quad\alpha_{1}(1+E)-E\leq B\delta^{*} (25)
CBδE(1+E)EBδ\displaystyle\Rightarrow\frac{C-B\delta^{*}}{E}(1+E)-E\leq B\delta^{*}\quad δδ2.\displaystyle\Rightarrow\quad\delta^{*}\geq\delta_{2}. (26)

Generalizing this for 1kD1\leq k\leq D, we obtain a general class of lower bounds as follows

αk1BδEαk1α1αk1\displaystyle\frac{\alpha_{k-1}-B\delta^{*}}{E}\leq\alpha_{k}\leq 1-\alpha_{1}-\cdots\alpha_{k-1} (27)
αk1(1+E)+Eαk2++Eα1EBδ\displaystyle\Rightarrow\qquad\alpha_{k-1}(1+E)+E\alpha_{k-2}+\cdots+E\alpha_{1}-E\leq B\delta^{*} (28)
αk2(1+E+E2)+E2αk3++E2α1E2B(1+2E)δ.\displaystyle\Rightarrow\alpha_{k-2}(1+E+E^{2})+E^{2}\alpha_{k-3}+\cdots+E^{2}\alpha_{1}-E^{2}\leq B(1+2E)\delta^{*}. (29)

Iterating this process, we obtain

α1j=0k1EjEk1Bδj=0k2Ej(j+1)\displaystyle\alpha_{1}\sum_{j=0}^{k-1}E^{j}-E^{k-1}\leq B\delta^{*}\sum_{j=0}^{k-2}E^{j}(j+1) (30)
(CBδE)j=0k1EjEk1Bδj=0k2Ej(j+1),\displaystyle\Rightarrow\left(\frac{C-B\delta^{*}}{E}\right)\sum_{j=0}^{k-1}E^{j}-E^{k-1}\leq B\delta^{*}\sum_{j=0}^{k-2}E^{j}(j+1), (31)

which upon rearranging implies δδk\delta^{*}\geq\delta_{k}. ∎

In general, none of these lower bounds dominates another. We will shortly elaborate on the relationship between δk\delta_{k} for k[D]k\in[D]. Before doing so, we first obtain another lower bound for δ\delta^{*}, termed a Type-II lower bound.

Lemma 2.

For the noise distribution given by Proposition 3, we have

δ\displaystyle\delta^{*} δD+11Bj=0D1Ej(Dj).\displaystyle\geq\delta_{D+1}\doteq\frac{1}{B\sum_{j=0}^{D-1}E^{j}(D-j)}. (32)
Proof.

First note that we can write from (17) that αDBδ\alpha_{D}\leq B\delta^{*}. We can extend this inequality for all αi\alpha_{i} via (18) as follows

αDkBδj=0kEj,\alpha_{D-k}\leq B\delta^{*}\sum_{j=0}^{k}E^{j}, (33)

for k[0:D1]k\in[0:D-1]. Since i=1Dαi=1\sum_{i=1}^{D}\alpha_{i}=1, the above inequality implies that

1\displaystyle 1 =k=0D1αDkBδk=0D1j=0kEj=Bδj=0D1k=jD1Ej=Bδj=0D1Ej(Dj),\displaystyle=\sum_{k=0}^{D-1}\alpha_{D-k}\leq B\delta^{*}\sum_{k=0}^{D-1}\sum_{j=0}^{k}E^{j}=B\delta^{*}\sum_{j=0}^{D-1}\sum_{k=j}^{D-1}E^{j}=B\delta^{*}\sum_{j=0}^{D-1}E^{j}(D-j),

from which the result follows by a arrangement. ∎

Putting lemmas 1 and 2 together, the following theorem is straightforward.

Theorem 1.

For the noise distribution given by Proposition 3, we have

δδ~,\delta^{*}\geq\tilde{\delta},

where δ~maxk[D+1]δk\tilde{\delta}\triangleq\max_{k\in[D+1]}{\delta_{k}} and δk\delta_{k}’s were defined in (23) for k[D]k\in[D] and in (32) for k=D+1k=D+1.


To show that this lower bound is in fact tight, we need to carefully investigate the dynamic of δk\delta_{k}’s. To do so, we define crossover values CkC_{k} as

Ck\displaystyle C_{k} j=0kEjj=0k1Ej(kj),k[D].\displaystyle\doteq\frac{\sum_{j=0}^{k}E^{j}}{\sum_{j=0}^{k-1}E^{j}(k-j)},\quad k\in[D]. (34)

For any k[D]k\in[D], the relational order between two consecutive lower bounds δk\delta_{k} and δk+1\delta_{k+1} is uniquely determined by the relation of the parameter C2ηη¯C\doteq\frac{2\eta}{\bar{\eta}} with the crossover value CkC_{k} as prescribed in the following lemma. This, together with strictly decreasing behavior of the sequence {Ck}k=1D\{C_{k}\}_{k=1}^{D} (to be proved in Lemma 4) will ensure that we can find the maximum among all δk\delta_{k}, k[D+1]k\in[D+1] analytically.

Lemma 3.

We have δkδk+1\delta_{k}\geq\delta_{k+1} for k[D]k\in[D] if and only if η\eta is such that C2ηη¯CkC\doteq\frac{2\eta}{\bar{\eta}}\geq C_{k}.

Proof.

First consider k=1k=1 and note that

δ1δ20CEC+CEE21+2EC1+E=C1.\displaystyle\delta_{1}-\delta_{2}\geq 0\quad\Leftrightarrow\quad C-E\geq\frac{C+CE-E^{2}}{1+2E}\quad\Leftrightarrow\quad C\geq 1+E=C_{1}. (35)

For k[2:D1]k\in[2:D-1], it follows that δkδk+1\delta_{k}\geq\delta_{k+1} if and only if

Cj=0k1EjEkBj=0k1Ej(j+1)\displaystyle\frac{C\sum_{j=0}^{k-1}E^{j}-E^{k}}{B\sum_{j=0}^{k-1}E^{j}(j+1)} Cj=0kEjEk+1Bj=0kEj(j+1).\displaystyle\geq\frac{C\sum_{j=0}^{k}E^{j}-E^{k+1}}{B\sum_{j=0}^{k}E^{j}(j+1)}.

After a straightforward algebraic manipulation, we can show that it is equivalent to

Cj=0kEjj=0k1Ej(kj)=Ck.C\geq\frac{\sum_{j=0}^{k}E^{j}}{\sum_{j=0}^{k-1}E^{j}(k-j)}=C_{k}.

Last, we have δDδD+1\delta_{D}\geq\delta_{D+1} if and only if

Cj=0D1EjEDBj=0D1Ej(j+1)1Bj=0D1Ej(Dj),\displaystyle\frac{C\sum_{j=0}^{D-1}E^{j}-E^{D}}{B\sum_{j=0}^{D-1}E^{j}(j+1)}\geq\frac{1}{B\sum_{j=0}^{D-1}E^{j}(D-j)},

which is in turn equivalent to

C(j=0D1Ej)(j=0D1Ej(Dj))ED(j=0D1Ej(Dj))+j=0D1Ej(j+1)\displaystyle C\left(\sum_{j=0}^{D-1}E^{j}\right)\left(\sum_{j=0}^{D-1}E^{j}(D-j)\right)\geq E^{D}\left(\sum_{j=0}^{D-1}E^{j}(D-j)\right)+\sum_{j=0}^{D-1}E^{j}(j+1)
=(j=0D1Ej)(j=0DEj).\displaystyle=\left(\sum_{j=0}^{D-1}E^{j}\right)\left(\sum_{j=0}^{D}E^{j}\right).

Hence, we conclude δDδD+1\delta_{D}\geq\delta_{D+1} if and only if

Cj=0D1Ejj=0D1Ej(Dj)=CD.C\geq\frac{\sum_{j=0}^{D-1}E^{j}}{\sum_{j=0}^{D-1}E^{j}(D-j)}=C_{D}.

Lemma 4.

The sequence {Ck}k=1D\{C_{k}\}_{k=1}^{D} is strictly decreasing.

Proof.

This can be easily verified by writing the difference between CkC_{k} and Ck+1C_{k+1} and checking that

CkCk+1=j=0kEj(j+1)(j=0k1Ej(kj))(j=0kEj(k+1j))>0,k[D].\displaystyle C_{k}-C_{k+1}=\frac{\sum_{j=0}^{k}E^{j}(j+1)}{(\sum_{j=0}^{k-1}E^{j}(k-j))(\sum_{j=0}^{k}E^{j}(k+1-j))}>0,\quad k\in[D]. (36)

For notational simplicity, define CD+10C_{D+1}\doteq 0 and C0C_{0}\doteq\infty. Thus, {Ck}k=0D+1\{C_{k}\}_{k=0}^{D+1} is still monotonically decreasing. Combining Lemma 3 with Lemma 4, we obtain the following result.

Theorem 2.

For the noise distribution given by Proposition 3, we have

δ~maxi[D+1]δi=δk,\tilde{\delta}\doteq\max_{i\in[D+1]}{\delta_{i}}=\delta_{k},

if and only if η\eta is such that Ck<CCk1C_{k}<C\leq C_{k-1}.

Proof.

To prove the only-if direction, note that due to strict monotonicity of {Ck}k=0D+1\{C_{k}\}_{k=0}^{D+1}, C>CkC>C_{k} implies C>Ck+1>>CDC>C_{k+1}>\cdots>C_{D}. Therefore, according to Lemma 3, C>CkC>C_{k} implies δk>δk+1>δk+2>>δD+1\delta_{k}>\delta_{k+1}>\delta_{k+2}>\dots>\delta_{D+1}. Similarly, CCk1C\leq C_{k-1} implies C<Ck2<<C1C<C_{k-2}<\cdots<C_{1}. Therefore, CCk1C\leq C_{k-1} implies δkδk1>δk2>>δ1\delta_{k}\geq\delta_{k-1}>\delta_{k-2}>\dots>\delta_{1}. In summary, Ck<CCk1C_{k}<C\leq C_{k-1} implies δ~=δk\tilde{\delta}=\delta_{k}. The forward direction can be similarly argued. ∎

Together with Theorem 1, this theorem implies that if Ck<CCk1C_{k}<C\leq C_{k-1}, then

δδk.\delta^{*}\geq\delta_{k}.

Next, we design a mechanism for which this lower bound is achieved.

III-B Achievability of Lower Bounds

In this section, we show that the lower bound δ~=maxi[D+1]δi\tilde{\delta}=\max_{i\in[D+1]}{\delta_{i}} on δ\delta^{*} is in fact achievable, i.e., there exists {αi}i=1D\{\alpha_{i}\}_{i=1}^{D} with αi0\alpha_{i}\geq 0 and i=1Dαi=1\sum_{i=1}^{D}\alpha_{i}=1 such that the mechanism (16) satisfies the constraints of the linear program (9) with δ=δ~\delta=\tilde{\delta}.

Definition 2 (Optimal α\alpha’s).

Set δ~=maxi[D+1]δi\tilde{\delta}=\max_{i\in[D+1]}{\delta_{i}} as given in Theorem 1. If η\eta is such that 0<CCD0<C\leq C_{D}, then δ~=δD+1\tilde{\delta}=\delta_{D+1} and we define αj\alpha_{j}^{*}’s as follows

αD\displaystyle\alpha_{D}^{*} =Bδ~=1=0D1E(D),\displaystyle=B\tilde{\delta}=\frac{1}{\sum_{\ell=0}^{D-1}E^{\ell}(D-\ell)}, (37)
αj\displaystyle\alpha_{j}^{*} =Eαj+1+Bδ~==0DjE=0D1E(D),j[1:D1].\displaystyle=E\alpha_{j+1}^{*}+B\tilde{\delta}=\frac{\sum_{\ell=0}^{D-j}E^{\ell}}{\sum_{\ell=0}^{D-1}E^{\ell}(D-\ell)},\qquad j\in[1:D-1]. (38)

If η\eta is such that C>CDC>C_{D}, then δ~=δk\tilde{\delta}=\delta_{k} for some k[D]k\in[D]. We then define αj\alpha_{j}^{*}’s for j[k]j\in[k] as

α1=CBδ~E,αj=αj1Bδ~E,j[2:k]\displaystyle\alpha_{1}^{*}=\frac{C-B\tilde{\delta}}{E},\qquad\alpha^{*}_{j}=\frac{\alpha_{j-1}^{*}-B\tilde{\delta}}{E},\qquad j\in[2:k] (39)

and set αj=0\alpha^{*}_{j}=0 for j[k+1:D]j\in[k+1:D].444In the proof of Theorem 3, we will show α1>0\alpha_{1}^{*}>0, αj0\alpha_{j}^{*}\geq 0 for j[2:k]j\in[2:k] in (39) and that j=1Dαj=1\sum_{j=1}^{D}\alpha_{j}^{*}=1 as desired.

Now we are ready for the main result of this section.

Theorem 3.

The mechanism (16) with coefficients {αi}\{\alpha_{i}^{*}\} defined in Definition 2 satisfies the constraints of the linear program (9) with δ=δ~\delta=\tilde{\delta}, defined in Theorem 1. In particular,

δ=δ~.\delta^{*}=\tilde{\delta}.
Proof.

First note that once we prove that δ~\tilde{\delta} is achievable, Theorem 1 implies that δ=δ~\delta^{*}=\tilde{\delta}. Thus, we only need to prove the achievability of δ~\tilde{\delta}. Note further that {αj}\{\alpha^{*}_{j}\} satisfy the inequalities in Proposition 4, with δ\delta^{*} replaced by δ~\tilde{\delta}, with equality. Therefore, we only need to show that {αi}\{\alpha^{*}_{i}\} are valid in the sense that they are non-negative and sum to 1.

Assume C>CDC>C_{D} in general and in particular, we have Ck<CCk1C_{k}<C\leq C_{k-1}, for some k[D]k\in[D] (recall that C0=C_{0}=\infty). It follows from Theorem 2 that δ~=δk\tilde{\delta}=\delta_{k}, where δk\delta_{k} is a Type-I lower bound given in (23). First, we note that according to (39)

α1=CBδ~E=C=0k2E(+1)+Ek1=0k1E(+1)>0.\displaystyle\alpha_{1}^{*}=\frac{C-B\tilde{\delta}}{E}=\frac{C\sum_{\ell=0}^{k-2}E^{\ell}(\ell+1)+E^{k-1}}{\sum_{\ell=0}^{k-1}E^{\ell}(\ell+1)}>0. (40)

Now consider αk\alpha_{k}^{*} in (39). Applying (39) iteratively, we obtain

αk\displaystyle\alpha_{k}^{*} =αk1Bδ~E=CBδk=0k1EEk\displaystyle=\frac{\alpha_{k-1}^{*}-B\tilde{\delta}}{E}=\frac{C-B\delta_{k}\sum_{\ell=0}^{k-1}E^{\ell}}{E^{k}} (41)
==0k1EC=0k2E(k1)=0k1E(+1)\displaystyle=\frac{\sum_{\ell=0}^{k-1}E^{\ell}-C\sum_{\ell=0}^{k-2}E^{\ell}(k-1-\ell)}{{\sum_{\ell=0}^{k-1}E^{\ell}(\ell+1)}} (42)
=0k1ECk1=0k2E(k1)=0k1E(+1)=0,\displaystyle\geq\frac{\sum_{\ell=0}^{k-1}E^{\ell}-C_{k-1}\sum_{\ell=0}^{k-2}E^{\ell}(k-1-\ell)}{{\sum_{\ell=0}^{k-1}E^{\ell}(\ell+1)}}=0, (43)

where in the last two steps we have used the fact that CCk1C\leq C_{k-1} and the definition of Ck1C_{k-1} in (34). Note that the definition of {αi}\{\alpha_{i}^{*}\} in (39) together with555Note that from (32) we conclude δD+1>0\delta_{D+1}>0. Therefore, δ~=maxiδi>0\tilde{\delta}=\max_{i}\delta_{i}>0. δ~>0\tilde{\delta}>0 implies that αi<αj\alpha^{*}_{i}<\alpha^{*}_{j} for 2j<iD2\leq j<i\leq D. In particular, we have αkj>αk0\alpha_{k-j}^{*}>\alpha_{k}^{*}\geq 0 for j[k1]j\in[k-1]. Note that

i=1Dαi=i=1kαi=Cj=0k1EjBδkj=1kjEj1Ek=1,\displaystyle\sum_{i=1}^{D}\alpha_{i}^{*}=\sum_{i=1}^{k}\alpha_{i}^{*}=\frac{C\sum_{j=0}^{k-1}E^{j}-B\delta_{k}\sum_{j=1}^{k}jE^{j-1}}{E^{k}}=1, (44)

where we use the fact that αj=0\alpha_{j}^{*}=0 for j[k+1:D]j\in[k+1:D].

Finally, if CCDC\leq C_{D} then from Theorem 2, we conclude δ=δD+1\delta^{*}=\delta_{D+1}, which is the Type-II lower bound on δ\delta^{*}. In this case, {αi}\{\alpha^{*}_{i}\} are given in (37) and (38). A similar argument as above shows that 0αi10\leq\alpha^{*}_{i}\leq 1 for each i[D]i\in[D] and also iαi=1\sum_{i}\alpha^{*}_{i}=1.

This theorem demonstrates that {αi}\{\alpha^{*}_{i}\}, defined in Definition 2, constructs the “optimal” mechanism when plugging into (16) or equivalently the optimal noise distribution when plugging into (14). Fig. 1 shows such optimal noise distribution for ε=2.18{\varepsilon}=2.18 (E=8.8463E=8.8463), η=0.8\eta=0.8 (C=8C=8), and D=6D=6 for z0z\geq 0 (noting the symmetry for z<0z<0). Note that the minimum true count is assumed to be nD=6n\geq D=6, which seems a reasonable assumption in large datasets with hundreds of participants. Using (34), we find that 7.8867=C3<CC2=8.12297.8867=C_{3}<C\leq C_{2}=8.1229. Therefore, δ~=δ3=0.0049\tilde{\delta}=\delta_{3}=0.0049 according to Theorem 2. It follows from (39) that only α1\alpha_{1}^{*}, α2\alpha_{2}^{*}, and α3\alpha_{3}^{*} are non-zero. Consequently, the noise distribution is in fact supported on [3:3][-3:3], instead of what we originally required in P1, i.e., [6:6][-6:6].

Using Equation (1) in [12], we also plot in Fig. 1, discrete Gaussian distribution, denoted by pG,Z(z)p_{\text{G},Z}(z), for z0z\geq 0. We set its variance identical to that of our noise distribution, i.e., σ2=η¯i=1Dαii2\sigma^{2}=\bar{\eta}\sum_{i=1}^{D}\alpha_{i}^{*}i^{2}. It is worth noting that discrete Gaussian distribution is not compactly supported; however, the probability that it assigns to points outside [D:D][-D:D] is negligible and is not shown in Fig. 1. While the two distributions may look somewhat similar, they have important differences that substantially impact their privacy performance. Notice, for instance, that pG,Z(±1)=0.11685p_{\text{G},Z}(\pm 1)=0.11685 and pG,Z(±2)=0.000416p_{\text{G},Z}(\pm 2)=0.000416, and hence pY(n+2|n+1)pY(n+2|n)=pY(n1|n)pY(n1|n+1)\frac{p_{Y}(n+2|n+1)}{p_{Y}(n+2|n)}=\frac{p_{Y}(n-1|n)}{p_{Y}(n-1|n+1)} can be as large as about 282. Note that in our proposed distribution, pZ(±1)=η¯2α1=0.08987p_{Z}(\pm 1)=\frac{\bar{\eta}}{2}\alpha_{1}^{*}=0.08987 and pZ(±2)=η¯2α2=0.00960p_{Z}(\pm 2)=\frac{\bar{\eta}}{2}\alpha_{2}^{*}=0.00960, and hence pY(n+2|n+1)pY(n+2|n)=pY(n1|n)pY(n1|n+1)=α1α2=0.089870.00960=9.3617\frac{p_{Y}(n+2|n+1)}{p_{Y}(n+2|n)}=\frac{p_{Y}(n-1|n)}{p_{Y}(n-1|n+1)}=\frac{\alpha^{*}_{1}}{\alpha^{*}_{2}}=\frac{0.08987\cdots}{0.00960\cdots}=9.3617. Consequently, in the discrete Gaussian mechanism the smallest ε{\varepsilon} such that pY(n1|n)eεpY(n1|n+1)+δ~p_{Y}(n-1|n)\leq e^{\varepsilon}p_{Y}(n-1|n+1)+\tilde{\delta} for δ~=0.0049\tilde{\delta}=0.0049 is ε=5.6{\varepsilon}=5.6, while for our mechanism ε=2.18{\varepsilon}=2.18 is sufficient. This indicates that the optimally-chosen coefficient {αi}\{\alpha_{i}\} in our setting improves the privacy guarantee of discrete Gaussian mechanism. In the next section, we provide more detailed comparison between there two mechanisms.

Refer to caption
Figure 1: The noise distribution (14) with {αi}\{\alpha_{i}\} being chosen according to Definition 2 is compared with the discrete Gaussian noise [12] with the identical variance. Here, we assume for ε=2.18{\varepsilon}=2.18, η=0.8\eta=0.8, and D=6D=6. The more ‘graceful‘ transition in our noise distribution from z=1z=1 to z=2z=2 results in an improved privacy performance.
Remark 2.

As mentioned in Remark 1, we restricted the definition of differential privacy in Definition 1 to the singular events. As a result, the parameters ε{\varepsilon} and δ\delta computed in this section do not correspond to the DP parameters. To compute the relationship between these parameters, recall that the mechanism (16) is (ε,δ)({\varepsilon},\delta)-DP if for any event S[N+D]S\subset[N+D]

pY|n(S)eεpY|n+1(S)+δandpY|n(S)eεpY|n1(S)+δ,p_{Y|n}(S)\leq e^{\varepsilon}p_{Y|n+1}(S)+\delta\qquad\text{and}\qquad p_{Y|n}(S)\leq e^{\varepsilon}p_{Y|n-1}(S)+\delta,

for all possible query responses n[N]n\in[N]. Since pY|n(S)=ySpY|n(y)p_{Y|n}(S)=\sum_{y\in S}p_{Y|n}(y), our analysis in this section implies that for any S[N+D]S\subset[N+D]

pY|n(S)eεpY|n+1(S)\displaystyle p_{Y|n}(S)-e^{\varepsilon}p_{Y|n+1}(S) =yS[pY|n(y)eεpY|n+1(y)]\displaystyle=\sum_{y\in S}\left[p_{Y|n}(y)-e^{\varepsilon}p_{Y|n+1}(y)\right]
ySsupp(pY|n)[pY|n(y)eεpY|n+1(y)]\displaystyle\leq\sum_{y\in S\cap\text{supp}(p_{Y|n})}\left[p_{Y|n}(y)-e^{\varepsilon}p_{Y|n+1}(y)\right]
|Ssupp(pY|n)|maxy[N+D][pY|n(y)eεpY|n+1(y)]\displaystyle\leq|S\cap\text{supp}(p_{Y|n})|\max_{y\in[N+D]}\left[p_{Y|n}(y)-e^{\varepsilon}p_{Y|n+1}(y)\right]
(2D+1)maxy[N+D][pY|n(y)eεpY|n+1(y)]\displaystyle\leq(2D+1)\max_{y\in[N+D]}\left[p_{Y|n}(y)-e^{\varepsilon}p_{Y|n+1}(y)\right]
(2D+1)δ,\displaystyle\leq(2D+1)\delta^{*},

where supp(pY|n)\text{supp}(p_{Y|n}) denotes the set of y[N+D]y\in[N+D] with pY|n(y)>0p_{Y|n}(y)>0 and a closed-form expression for δ\delta^{*} was given in Theorem 3. The same argument shows that pY|n(S)eεpY|n1(S)(2D+1)δp_{Y|n}(S)-e^{\varepsilon}p_{Y|n-1}(S)\leq(2D+1)\delta^{*}. This observation indicates that the mechanism (16) is (ε,min{1,(2D+1)δ})({\varepsilon},\min\{1,(2D+1)\delta^{*}\})-DP when {αi}\{\alpha_{i}\} is chosen according to Definition 2.

IV Numerical Results

We begin this section by numerically computing the DP parameters ε{\varepsilon} and (2D+1)δ(2D+1)\delta^{*} of our mechanism according to Theorem 3.666We mostly consider practical parameter ranges resulting in (2D+1)δ1(2D+1)\delta^{*}\ll 1. We thus write (2D+1)δ(2D+1)\delta^{*} instead of min{1,(2D+1)δ}\min\{1,(2D+1)\delta^{*}\}. In Fig. 2, we assume η=0.5\eta=0.5 and plot the (2D+1)δ(2D+1)\delta^{*} in terms of ε{\varepsilon} for different values of DD. It is clear that (2D+1)δ(2D+1)\delta^{*} is no greater than 0.0010.001 for D=8D=8 and ε>1.1{\varepsilon}>1.1. It is worth noting that while we reveal the true count with probability 0.50.5, from Definition 2, the mechanism’s output lies in an acceptable range of [n3:n+3][n-3:n+3] with a much higher probability η+η¯(α1+α2+α3)=0.9945\eta+\bar{\eta}(\alpha_{1}^{*}+\alpha_{2}^{*}+\alpha_{3}^{*})=0.9945 for ε=1.5{\varepsilon}=1.5.

Refer to caption
Figure 2: DP parameters (2D+1)δ(2D+1)\delta^{*} versus ε{\varepsilon} for η=0.5\eta=0.5 and different values of DD.

In Fig. 3, we illustrate the relationship between the DP parameter (2D+1)δ(2D+1)\delta^{*} and η\eta for different values of ε{\varepsilon} and DD. Quite predictably, for given DD and ε{\varepsilon}, there exists a threshold for η\eta above which (2D+1)δ(2D+1)\delta^{*} grows fast to one; thus indicating a trade-off between utility (reliability) and privacy. To strike a good balance between the reliability and privacy, one may choose the the largest value of η\eta just before the ‘knee’ phenomenon occurs. For instance, assuming D=8D=8, our mechanism provides (1.1,103)(1.1,10^{-3})-DP and (2.2,5×107)(2.2,5\times 10^{-7})-DP guarantee while presenting η=0.5\eta=0.5 and η=0.8\eta=0.8, respectively. We remark that ε=1.1{\varepsilon}=1.1 and ε=2.2{\varepsilon}=2.2 were numerically chosen such that the knee phenomenon occurs slightly after η=0.5\eta=0.5 and η=0.8\eta=0.8, respectively.

Refer to caption
Refer to caption
Figure 3: DP parameter (2D+1)δ(2D+1)\delta^{*} versus η\eta for ε=1.1{\varepsilon}=1.1 (left) and ε=2.2{\varepsilon}=2.2 (right) and different values of DD.

Next, we compare our mechanism with the recently proposed discrete Gaussian mechanism [12] in terms of the utility-privacy performance. There may be different ways for conducting such comparison. To have a fair comparison, we take the following steps:

  1. 1.

    Given fixed values of η(0,1)\eta\in(0,1), DD\in\mathbb{N}, and ε0{\varepsilon}\geq 0, we compute the optimal coefficient {αi}\{\alpha^{*}_{i}\} (according to Definition 2) and the corresponding DP parameter (2D+1)δ(2D+1)\delta^{*} (according to Theorem 3 and Remark 2). We then compute σ2(η,D,ε)\sigma^{2}(\eta,D,{\varepsilon}) the noise variance given by

    σ2(η,D,ε)η¯i=1Dαii2.\sigma^{2}(\eta,D,{\varepsilon})\doteq\bar{\eta}\sum_{i=1}^{D}\alpha_{i}^{*}i^{2}.
  2. 2.

    We generate a discrete Gaussian probability mass function (pmf) according to Equation (1) in [12] with the variance σ2(η,D,ε)\sigma^{2}(\eta,D,{\varepsilon}). This dictates that the discrete Gaussian mechanism has the same utility as our mechanism. We refer to the resulting distribution as pG,Zp_{\text{G},Z}.

  3. 3.

    Let εG{\varepsilon}_{\text{G}} and δG\delta_{\text{G}} be the DP parameters of the discrete Gaussian mechanism. It is proved in [12, Theorem 7] that

    δG=pG,Z(Z>εGσ20.5)eεGpG,Z(Z>εGσ2+0.5),\displaystyle\delta_{\text{G}}=p_{\text{G},Z}\left(Z>{\varepsilon}_{\text{G}}\sigma^{2}-0.5\right)-e^{{\varepsilon}_{\text{G}}}p_{\text{G},Z}\left(Z>{\varepsilon}_{\text{G}}\sigma^{2}+0.5\right), (45)

    where σ2\sigma^{2} is the variance of the discrete Gaussian noise added to the true count. By replacing σ2\sigma^{2} by σ2(η,D,ε)\sigma^{2}(\eta,D,{\varepsilon}), we make use of this expression to compare the discrete Gaussian mechanism with our mechanism in terms of privacy parameters.

Refer to caption
Figure 4: Comparison of (ε,δ)({\varepsilon},\delta)-DP performance for η=0.5\eta=0.5 and D=6D=6 in our proposed mechanism with that of a discrete Gaussian mechanism with the same variance σ2(η,D,ε)\sigma^{2}(\eta,D,{\varepsilon}). In our mechanism, for a given ε0{\varepsilon}\geq 0, we compute the optimal coefficient {αi}\{\alpha^{*}_{i}\} (according to Definition 2) and the corresponding DP parameter (2D+1)δ(2D+1)\delta^{*} (according to Theorem 3 and Remark 2). For the discrete Gaussian mechanism, we use (45) to compute δG\delta_{G}.

As depicted in Fig 4, the (ε,δ)({\varepsilon},\delta)-DP performance of our proposed scheme is superior to that of the discrete Gaussian mechanism with the same variance for a wide range of privacy parameters (both mechanisms perform rather poorly for ε<1.1{\varepsilon}<1.1, resulting in quite high δ\delta). Many other experiments were conducted and they all showed trends similar to what is shown in Fig. 4.

V Conclusion

In this work, we explicitly constructed a privacy-preserving mechanism for responding to integer-valued queries to a dataset containing sensitive attributes. This mechanism is noise-additive but, unlike many other noise-additive private mechanisms, it adds an integer-valued noise to the true count. The noise distribution is parameterized by η(0,1)\eta\in(0,1) which specifies the allowable probability of error in responding the true count, thereby balancing the privacy-utility trade-off. We proved that this mechanism is (ε,δ)({\varepsilon},\delta)-differentially private where both ε{\varepsilon} and δ\delta depend on noise support size and η\eta. This mechanism is contrasted with the recently proposed discrete Gaussian mechanism [12] which adds discretized Gaussian noise (with infinite support) to the true count. Our numerical findings indicate that the resulting ε{\varepsilon} and δ\delta are tighter than those obtained in discrete Gaussian mechanism for many practical range of η\eta.

We conclude this work with a note on future direction. In this work, we assume that there is a single query to a dataset. However, in many practical scenarios there are several queries to a single dataset, each of which might depend on the previous ones. Provided that each query is responded by a our private mechanism, it is essential to determine how privacy degrades as the number of queries increases. We believe that the new technique propose in [15] might be helpful for this purpose.

References