This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Bounded-Loss Private Prediction Markets

Rafael Frongillo
Colorado Boulder
raf@colorado.edu
&Bo Waggoner
Microsoft Research
bwag@colorado.edu
Abstract

Prior work has investigated variations of prediction markets that preserve participants’ (differential) privacy, which formed the basis of useful mechanisms for purchasing data for machine learning objectives. Such markets required potentially unlimited financial subsidy, however, making them impractical. In this work, we design an adaptively-growing prediction market with a bounded financial subsidy, while achieving privacy, incentives to produce accurate predictions, and precision in the sense that market prices are not heavily impacted by the added privacy-preserving noise. We briefly discuss how our mechanism can extend to the data-purchasing setting, and its relationship to traditional learning algorithms.

1 Introduction

In a prediction market, a platform maintains a prediction (usually a probability distribution or an expectation) of a future random variable such as an election outcome. Participants’ trades of financial securities tied to this event are translated into updates to the prediction. Prediction markets, designed to aggregate information from participants, have gained a substantial following in the machine learning literature. One reason is the overlap in goals (predicting future outcomes) as well as techniques (convex analysis, Bregman divergences), even at a deep level: the form of market updates in standard automated market makers have been shown to mimic standard online learning or optimization algorithms in many settings Frongillo et al. (2012); Abernethy et al. (2013, 2014a); Frongillo and Reid (2015). Beyond this research-level bridge, recent papers have suggested prediction market mechanisms as a way of crowdsourcing data or algorithms for machine learning, usually by providing incentives for participants to repeatedly update a centralized hypothesis or prediction Abernethy and Frongillo (2011); Waggoner et al. (2015).

One recently-proposed mechanism to purchase data or hypotheses from participants is that of Waggoner, et al. Waggoner et al. (2015), in which participants submit updates to a centralized market maker, either by directly altering the hypothesis, or in the form of submitted data; both are interpreted as buying or selling shares in a market, paying off according to a set of holdout data that is revealed after the close of the market. The authors then show how to preserve differential privacy for participants, meaning that the content of any individual update is protected, as well as natural accuracy and incentive guarantees.

One important drawback of Waggoner, et al. Waggoner et al. (2015), however, is the lack of a bounded worst-case loss guarantee: as the number of participants grows, the possible financial liability of the mechanism grows without bound. In fact, their mechanism cannot achieve a bounded worst-case loss without giving up privacy guarantees. Subsequently, Cummings, et al. Cummings et al. (2016) show that all differentially-private prediction markets of the form proposed in Waggoner et al. (2015) must suffer from unbounded financial loss in the worst case. Intuitively, one could interpret this negative result as saying that the randomness of the mechanism, which must be introduced to preserve privacy, also creates arbitrage opportunities for participants: by betting against the noise, they collectively expect to make an unbounded profit from the market maker. Nevertheless, Cummings, et al. leave open the possibility that mechanisms outside the mold of Waggoner, et al. could achieve both privacy and a bounded worst-case loss.

In this paper, we give such a mechanism: the first private prediction market framework with a bounded worst-case loss. When applied to the crowdsourcing problems stated above, this now allows the mechanism designer to maintain a fixed budget. Our construction and proof proceeds in two steps.

We first show that by adding a small transaction fee to the mechanism of Waggoner et al. (2015), one can eliminate financial loss due to arbitrage while maintaining the other desirable properties of the market. The key idea is that a carefully-chosen transaction fee can make each trader subsidize (in expectation) any arbitrage that may result from the noise preserving her privacy. Unless prices already match her beliefs quite closely, however, she still expects to make a profit by paying the fee and participating. We view this as a positive result both conceptually—it shows that arbitrage opportunities are not an insurmountable obstacle to private markets—and technically—the designer budget grows very slowly, only O((logT)2)O((\log T)^{2}), with the number of participants TT.

Nonetheless, this first mechanism is still not completely satisfactory, as the budget is superconstant in TT, and TT must be known in advance. This difficulty arises not from arbitrage, but (apparently) a deeper constraint imposed by privacy that forces the market to be large relative to the participants. Our second and main result overcomes this final hurdle. We construct a sequence of adaptively-growing markets that are syntactically similar to the “doubling trick” in online learning. The key idea is that, in the market from our first result, only about (logT)2(\log T)^{2} of the TT participants can be informational traders; after this point, additional participants do not cost the designer any more budget, yet their transaction fees can raise significant funds. So if the end of a stage is reached, the market activity has actually generated a surplus which subsidizes the initial portion of the next stage of the market.

2 Setting

In a cost-function based prediction market, there is an observable future outcome ZZ taking values in a set 𝒵\mathcal{Z}. The goal is to predict the expectation of a random variable ϕ:𝒵d\phi:\mathcal{Z}\to\mathbb{R}^{d}. We assume ϕ\phi is a bounded random variable, as otherwise prediction markets with bounded financial loss are not possible. Participants will buy from the market contracts, each parameterized by a vector rdr\in\mathbb{R}^{d}. The contract represents a promise for the market to pay the owner rϕ(Z)r\cdot\phi(Z) when ZZ is observed. Adopting standard financial terminology, in our model there are dd securities j=1,,dj=1,\dots,d, and the owner of a share in security jj will receive a payoff of ϕ(Z)j\phi(Z)_{j}, that is, the jjth component of the random variable. Thus a contract rdr\in\mathbb{R}^{d} contains rjr_{j} shares of security jj and pays off j=1drjϕ(Z)j=rϕ(Z)\sum_{j=1}^{d}r_{j}\phi(Z)_{j}=r\cdot\phi(Z). Note that rj<0r_{j}<0, or “short selling” security jj, is allowed.

The market maintains a market state qtdq^{t}\in\mathbb{R}^{d} at time t=0,,Tt=0,\dots,T, with q0=0q^{0}=0. Each trader t=1,,Tt=1,\dots,T arrives sequentially and purchases a contract dqtddq^{t}\in\mathbb{R}^{d}, and the market state is updated to qt=qt1+dqtq^{t}=q^{t-1}+dq^{t}. In other words, qt=s=1tdqsq^{t}=\sum_{s=1}^{t}dq^{s}, the sum of all contracts purchased up to time tt. The price paid by each participant is determined by a convex cost function C:dC:\mathbb{R}^{d}\to\mathbb{R}. Intuitively, CC maps qtq^{t} to the total price paid by all agents so far, C(qt)C(q^{t}). Thus, participant tt making trade dqtdq^{t} when the current state is qt1q^{t-1} pays C(qt1+dqt)C(qt1)C(q^{t-1}+dq^{t})-C(q^{t-1}). Notice that the instantaneous prices pt=C(qt)p^{t}=\nabla C(q^{t}) represent the current price per unit of infinitesimal purchases, with the jjth coordinate representing the current price per share of the jjth security.

The prices C(q)\nabla C(q) are interpreted as predictions of 𝔼ϕ(Z)\operatorname*{\mathbb{E}}\phi(Z), as an agent who believes the jjth coordinate is too low will purchase shares in it, raising its price, and so on. This can be formalized through a learning lens: It is known (Abernethy et al., 2013) that agents in such a market maximize expected profit by minimizing an expected Bregman divergence between ϕ(Z)\phi(Z) and C(q)\nabla C(q); of course, it is known that C(q)=𝔼ϕ(Z)\nabla C(q)=\operatorname*{\mathbb{E}}\phi(Z) minimizes risk for any divergence-based loss Savage (1971); Banerjee et al. (2005); Abernethy and Frongillo (2012). (The Bregman divergence is that corresponding to CC^{*}, the convex conjugate of CC.)

Price Sensitivity.  The price sensitivity of a cost function CC is a measure of how quickly prices respond to trades, similar to “liquidity” discussed in Abernethy et al. (2013, 2014b) and earlier works. Formally, the price sensitivity λ\lambda of CC is the supremum of the operator norm of the Hessian of CC, with respect to the 1\ell_{1} norm.111For convenience we will assume CC is twice differentiable, though this is not necessary. In other words, if c=qq1c=\|q-q^{\prime}\|_{1} shares are purchased, then the change in prices C(q)C(q)1\|\nabla C(q)-\nabla C(q^{\prime})\|_{1} is at most λc\lambda c.

Price sensitivity is directly related to the worst-case loss guarantee of the market, as follows. Those familiar with market scoring rules may recall that with scoring rule SS, the loss can be bounded by (a constant times) the largest possible score. Hence, scaling SS by a factor 1λ\frac{1}{\lambda} immediately scales the loss bound by 1λ\frac{1}{\lambda} as well. Recall that SS is defined by a convex function GG, the convex conjugate of CC. Scaling SS by 1λ\frac{1}{\lambda} is equivalent to scaling GG by 1λ\frac{1}{\lambda}. By standard results in convex analysis, this is equivalent to transforming CC into Cλ(q)=1λC(λq)C_{\lambda}(q)=\frac{1}{\lambda}C\left(\lambda q\right), an operation known as the perspective transform. This in turn scales the price sensitivity by λ\lambda by the properties of the Hessian.

Price sensitivity is also related to the total number of trades required to change the prices in a market. If we assume each trade consists of at most one share in each security, then 1λ\frac{1}{\lambda} trades are necessary to shift the predictions to an arbitrary point from an arbitrary point.

Convention: normalized, scaled CC.  In the remainder of the paper, we will suppose that we start with some convex cost function C1C_{1} whose price sensitivity equals 11 and worst-case loss bounded by some constant B1B_{1}. Then, to obtain price sensitivity λ\lambda, we use the cost function C()=1λC1(λ)C(\cdot)=\frac{1}{\lambda}C_{1}(\lambda\cdot). As discussed above, CC has price sensitivity at most λ\lambda and a worst-case loss bound of B=B1/λB=B_{1}/\lambda. (This assumption is without loss of generality, as any cost function that guarantees a bounded worst-case loss can be scaled such that its price sensitivity is 11.)

2.1 Prior work

To achieve differential privacy for trades of a bounded size (which will be assumed), the general approach is to add random noise to the “true” market state qq and publish this noisy state q^\hat{q}. The privacy level thus determines how close q^\hat{q} is to qq. The distance from C(q^)\nabla C(\hat{q}) to C(q)\nabla C(q) is then controlled by the price sensitivity λ\lambda. For a fixed noise and privacy level, a smaller λ\lambda leads to small impact of noise on prices, meaning very good accuracy. However, decreasing λ\lambda does not come for free: the worst-case financial loss of to the market designer scales as 1/λ1/\lambda.

The market of Waggoner et al. (2015) adds controlled and correlated noise over time, in a manner similar to the “continual observation” technique of differential privacy. This reduces the influence of noise on accuracy to polylogarithmic in TT, the number of participants. Their main result for the prediction market setting studied here is as follows.

Theorem 1 (Waggoner et al. (2015)).

Assuming that all trades satisfy dqt11\|dq^{t}\|_{1}\leq 1, the private mechanism is ϵ\epsilon-differentially private in the trades dq1,,dqTdq^{1},\dots,dq^{T} with respect to the output q^1,,q^T\hat{q}^{1},\dots,\hat{q}^{T}. Further, to satisfy ptp^t1α\|p^{t}-\hat{p}^{t}\|_{1}\leq\alpha for all tt, except with probability γ\gamma, it suffices for the price sensitivity to be

λ=αϵ42dlogTln(2Td/γ).\displaystyle\lambda^{*}=\frac{\alpha~\epsilon}{4\sqrt{2}d\lceil\log T\rceil\ln(2Td/\gamma)}~. (1)

2.2 Our setting and desiderata

This paper builds on the work of Waggoner et al. (2015) to overcome the negative results of Cummings et al. (2016). Here, we formalize our setting and four desirable properties we hope to achieve.

Write a prediction market mechanism as a function MM taking inputs dq=dq1,,dqT\vec{dq}=dq^{1},\dots,dq^{T} and outputting a sequence of market states q^1,,q^T\hat{q}^{1},\dots,\hat{q}^{T}. Here q^t\hat{q}^{t} is thought of as a noisy version of qt=stdqsq^{t}=\sum_{s\leq t}dq^{s}. Each of these states is associated with a prediction p^t\hat{p}^{t} in the set of possible prices (expectations of ϕ\phi), while the state qtq^{t} is associated with the “true” underlying prediction ptp^{t}.

Definition 1 (Privacy).

MM satisfies (ϵ,δ)(\epsilon,\delta)-differential privacy if for all pairs of inputs dq,dq\vec{dq},\vec{dq^{\prime}} differing by only a single participants’ entry, and for all sets SS of possible outputs, Pr[M(dq)S]eϵPr[M(dq)S]+δ\Pr[M(\vec{dq})\in S]\leq e^{\epsilon}\Pr[M(\vec{dq^{\prime}})\in S]+\delta. If furthermore δ=0\delta=0, we say MM is ϵ\epsilon-differentially private.

Definition 2 (Precision).

MM has (α,γ)(\alpha,\gamma) precision if for all dq\vec{dq}, with probability 1γ1-\gamma, we have p^tpt1α\|\hat{p}^{t}-p^{t}\|_{1}\leq\alpha for all tt.

Definition 3 (Incentives).

MM has β\beta-incentive to participate if, for all beliefs p=𝔼ϕ(Z)p=\operatorname*{\mathbb{E}}\phi(Z), if at any point p^tp>β\|\hat{p}^{t}-p\|_{\infty}>\beta, then there exists a participation opportunity that makes a strictly positive profit in expectation with respect to pp.

For the budget guarantee, we must formalize the notion that participants may respond to the noise introduced by the mechanism. Following Cummings et al. (2016), let a trader strategy s=(s1,,sT)\vec{s}=(s^{1},\dots,s^{T}) where each sts^{t} is a possibly-randomized function of the form st(dq1,,dqt1;q^1,,q^t1)=dqts^{t}(dq^{1},\dots,dq^{t-1};\hat{q}^{1},\dots,\hat{q}^{t-1})=dq^{t}, i.e. a strategy taking the entire history prior to tt and outputting a trade dqtdq^{t}. Let L(M,s,z)L(M,\vec{s},z) be a random variable denoting the financial loss of the market MM against trader strategy s\vec{s} when Z=zZ=z, which for the mechanism described above is simply

L(M,s,z)=t=1T[C(q^t)C(q^t+dqt)dqtϕ(z)]\displaystyle L(M,\vec{s},z)=\sum_{t=1}^{T}\left[C(\hat{q}^{t})-C(\hat{q}^{t}+dq^{t})-dq^{t}\cdot\phi(z)\right] .

Definition 4.

MM guarantees designer budget BB if, for any trader strategy s\vec{s} and all zz, 𝔼L(M,s,z)B\operatorname*{\mathbb{E}}L(M,\vec{s},z)\leq B, where the expectation is over the randomness in MM and each sts^{t}.

3 Slowly-Growing Budget

The private market of Waggoner et al. (2015) causes unbounded loss for the market maker in two ways. The first is from traders betting against the random noise introduced to protect privacy. This is a key idea leveraged by Cummings et al. (2016) to show negative results for private markets. In this section, we show that a transaction fee can be chosen to exactly balance the expected profit from this type of arbitrage.222Intuitively, it is enough for the fee to cover arbitrage amounts in expectation, because a trader must pay the fee to trade before the random noise is drawn and any arbitrage opportunity is revealed. We will show that this fee is still small enough to allow for very accurate prices.333For instance, if the current price of a security is 0.490.49 and a trader believes the true price should be 0.500.50, she will purchase a share if the fee is c<0.01c<0.01. (For privacy, we limit each trade to a fixed size, say, one share.) This transaction fee restores the worst-case loss guarantee to the inverse of the price sensitivity, just as in a non-private market. The second way the market causes unbounded loss is to require price sensitivity to shrink as a function of TT; this is addressed in the next section.

We show that with this carefully-chosen fee, the market still achieves precision, incentive, and privacy guarantees, but now with a worst-case market maker loss of O((logT)2)O((\log T)^{2}), much improved over the naïve O(T)O(T) bound. This is viewed as a positive result because the worst-case loss is growing quite slowly in the total number of participants, and moreover matches the fundamental “informational” worst-case loss one expects with price sensitivity λ\lambda^{*}.

3.1 Mechanism and result

Here we recall the private market mechanism of Waggoner et al. (2015), adapted to the prediction market setting following Cummings et al. (2016). We will express the randomness of the mechanism in terms of a “noise trader” for both intuition and technical convenience. The market is defined by a cost function CC with price sensitivity λ\lambda, and parameters cc (transaction fee), ϵ\epsilon (privacy), α,γ\alpha,\gamma (precision), and TT (maximum number of participants). There is a special trader we call the noise trader who is controlled by the designer. All actions of the noise trader are hidden and known only by the designer. The designer publishes an initial market state q0=q^0=0q^{0}=\hat{q}^{0}=0. Let TT^{\prime} denote the actual number of arrivals, with TTT^{\prime}\leq T by assumption. Then, for t=1,,Tt=1,\dots,T^{\prime}:

  1. 1.

    Participant tt arrives, pays a fee of cc, and purchases bundle dqtdq^{t} with dqt11\|dq^{t}\|_{1}\leq 1. The payment is C(q^t+dqt)C(q^t)C(\hat{q}^{t}+dq^{t})-C(\hat{q}^{t}).

  2. 2.

    The noise trader purchases a randomly-chosen bundle ztz^{t}, called a noise trade, after selling off some subset {zt1,,ztk}\{z^{t_{1}},\ldots,z^{t_{k}}\} of previously purchased noise trades for ti<tt_{i}<t, according to a predetermined schedule described below. Letting wt=zti=1kztiw^{t}=z^{t}-\sum_{i=1}^{k}z^{t_{i}} denote this net noise bundle, the noise trader is thus charged C(q^t+dqt+wt)C(q^t+dqt)C(\hat{q}^{t}+dq^{t}+w^{t})-C(\hat{q}^{t}+dq^{t}).

  3. 3.

    The “true” market state is updated to qt=qt1+dqtq^{t}=q^{t-1}+dq^{t}, but is not revealed.

  4. 4.

    The noisy market state is updated to q^t=q^t1+dqt+wt\hat{q}^{t}=\hat{q}^{t-1}+dq^{t}+w^{t} and is published.

Finally, z𝒵z\in\mathcal{Z} is observed and each participant tt receives a payment dqtϕ(z)dq^{t}\cdot\phi(z). For the sake of budget analysis, we suppose that at the close of the market, the noise trader sells back all of her remaining bundles; letting wTw^{T^{\prime}} be the sum of these bundles, she is charged C(q^TwT)C(q^T)C(\hat{q}^{T^{\prime}}-w^{T^{\prime}})-C(\hat{q}^{T^{\prime}}).

Noise trades.  Each ztz^{t} is a dd-dimensional vector with each coordinate drawn from an independent Laplace distribution with parameter b=2logT/ϵb=2\lceil\log T\rceil/\epsilon. To determine which bundles zsz^{s} are sold at time tt, write t=2jmt=2^{j}m where mm is odd, and sell all bundles zsz^{s} purchased during the previous 2j12^{j-1} time steps which are not yet sold. Thus, the noise trader will sell bundles purchased at times s=t1,t2,t4,t8,,t2j1s=t-1,t-2,t-4,t-8,\dots,t-2^{j-1}; in particular, when tt is odd we have j=0j=0, so no previous bundles will be sold.

Budget.  The total loss of the market designer can now be written as the sum of three terms: the loss of the market maker, the loss of the noise trader, and the gain from transaction fees. By convention, the noise trader eventually sells back all bundles it purchases and is left with no shares remaining.

L(M,s,z)=t=1TC(q^t1)C(q^t1+dqt)+dqtϕ(z)net loss of market maker+t=1TC(q^t1+dqt)C(q^t)net loss of noise trader+cTfees.\displaystyle\!\!\!\!\!\!L(M,\vec{s},z)=\overbrace{\sum_{t=1}^{T^{\prime}}C(\hat{q}^{t-1})-C(\hat{q}^{t-1}\!\!+dq^{t})+dq^{t}\!\cdot\phi(z)}^{\text{net loss of market maker}}+\overbrace{\sum_{t=1}^{T^{\prime}}C(\hat{q}^{t-1}\!\!+dq^{t})-C(\hat{q}^{t})}^{\text{net loss of noise trader}}+\!\overbrace{cT^{\prime}\vphantom{\sum_{t=1}^{T^{\prime}}}}^{\text{fees}}\!\!.\!\! (2)

The main result of this section is as follows.

Theorem 2.

When each arriving participant pays a transaction fee c=αc=\alpha, the private market with any λλ\lambda\leq\lambda^{*} from eq. (1) satisfies ϵ\epsilon-differential privacy, (α,γ)(\alpha,\gamma)-precision, 2α2\alpha-incentive to trade, and budget bound B1λ\frac{B_{1}}{\lambda}, where B1B_{1} is the budget bound of the underlying cost function C1C_{1}.

3.2 Proof ideas: privacy, precision, incentives

The differential privacy and precision claims follow directly from the prior results, as nothing has changed to impact them. The incentive claim is not technically involved, but perhaps subtle: the transaction fee should be high enough to eliminate expected profit from arbitrage, yet low enough to allow for profit from information. The key point is that the transaction fee is a constant, but the farther the prices are from the trader’s belief, the more money she expects to make from a constant-sized trade. The transaction fee creates a ball of size 2α2\alpha around the current prices where, if one’s belief lies in that ball, then participation is not profitable.

We give most of the proof of the designer budget bound, with some claims deferred to the full version.

Lemma 1 (Budget bound).

The transaction-fee private market with any price sensitivity λλ\lambda\leq\lambda^{*} guarantees a designer budget bound of B1λ\frac{B_{1}}{\lambda}.

Proof.

Let cc be the transaction fee; we will later take c=αc=\alpha. Then the worst-case loss from eq. (2) is

WC(λ,T):=WC0(λ,T)+NTL(λ,T)Tc,WC(\lambda,T^{\prime}):=WC_{0}(\lambda,T^{\prime})+NTL(\lambda,T^{\prime})-T^{\prime}c~,

where WC0(λ,T)WC_{0}(\lambda,T^{\prime}) is the worst-case loss of a standard prediction market maker with parameter λ\lambda and TT^{\prime} participants, NTL(λ,T)NTL(\lambda,T^{\prime}) is the worst-case noise trader loss, and TcT^{\prime}c is the revenue from TT^{\prime} transaction fees of size cc each.

The worst-case loss of a standard prediction market maker is well-known; see e.g. Abernethy et al. (2013). By our normalization and definition of price sensitivity, we thus have WC0(λ,T)B1λWC_{0}(\lambda,T^{\prime})\leq\frac{B_{1}}{\lambda}.

To bound the noise trader loss NTL(λ,T)NTL(\lambda,T^{\prime}), we will consider each bundle ztz^{t} purchased by the noise trader. The idea is to bound the difference in price between the purchase and sale of ztz^{t}. For analysis, we suppose that at each tt, the noise trader first sells any previous bundles (e.g. at t=4t=4, first selling z3z^{3} and then selling z2z^{2}), and then purchases ztz^{t}.

Now let b(t)b(t) be the largest power of 22 that divides tt. Let qbuytq_{\text{buy}}^{t} and qselltq_{\text{sell}}^{t} be the market state just before the noise trader purchases ztz^{t} and just after she sells ztz^{t}, respectively.

Claim 1.

For each tt, exactly b(t)b(t) traders arrive between the purchase and the sale of bundle ztz^{t}; furthermore, qselltqbuytq_{\text{sell}}^{t}-q_{\text{buy}}^{t} is exactly equal to the sum of these participants’ trades.

For example, suppose tt is odd. Then only one participant arrives between the purchase and sale of ztz^{t}. Furthermore, ztz^{t} is the last bundle purchased by the noise trader at time tt and is the first sold at time t+1t+1, so the difference in market state is exactly ztz^{t} plus that participant’s trade.

Claim 2.

If the noise trader purchases and later sells ztz^{t}, then her net loss in expectation over ztz^{t} (but for any trader behavior in response to ztz^{t}), is at most λb(t)K\lambda b(t)K where K=𝔼zt2K=\operatorname*{\mathbb{E}}\|z^{t}\|_{2}.

We now sum over all bundles ztz^{t} purchased by the noise trader, i.e. at time steps 1,,T1,\dots,T^{\prime}. Recall that the noise trader sells back every bundle ztz^{t} she purchases. Thus, her total payoff is the sum over tt of the difference in price at which she buys ztz^{t} and price at which she sells it. For each j=0,,logT1j=0,\dots,\log T^{\prime}-1, there are 2j2^{j} different steps tt with b(t)=T/2j+1b(t)=T^{\prime}/2^{j+1}. The total loss is thus,

NTL(λ,T)\displaystyle NTL(\lambda,T^{\prime}) j=0logT12jT2j+1λK=TlogT2λK.\displaystyle\leq\sum_{j=0}^{\log T^{\prime}-1}2^{j}\frac{T^{\prime}}{2^{j+1}}\lambda K=\frac{T^{\prime}\log T^{\prime}}{2}\lambda K~. (3)

Note that if the noise trader has some noise bundles left over after the final participant, we suppose she immediately sells all remaining bundles back to the market maker in reverse order of purchase.

Putting eq. (3) together with the above bound on WC0WC_{0} gives

WC(λ,T)\displaystyle WC(\lambda,T^{\prime}) WC0(λ,T)+TlogTλKTcB1λ+T(KlogTλc),\displaystyle\leq WC_{0}(\lambda,T^{\prime})+T^{\prime}\log T^{\prime}\lambda K-T^{\prime}c\leq\frac{B_{1}}{\lambda}+T^{\prime}\left(K\log T^{\prime}\lambda-c\right)~, (4)

which is in turn at most B1/λB_{1}/\lambda if we choose λ\lambda and the transaction fee cc such that cKlogTλc\geq K\log T\lambda. In other words, we take λc/KlogT\lambda\leq c/K\log T.

Finally, we can bound K=𝔼zt2K=\operatorname*{\mathbb{E}}\|z^{t}\|_{2} from Claim 2 as follows: for each tt, the components of the dd-dimensional vector ztz^{t} are each independent Lap(b)\mathrm{Lap}(b) variables with b=2logT/ϵb=2\lceil\log T\rceil/\epsilon. By concavity of \sqrt{\cdot}, we have

K=𝔼i=1dzt(i)2i𝔼zt(i)2=dVar(Lap(b))=2db2=22dlogTϵ.\displaystyle K=\operatorname*{\mathbb{E}}\sqrt{\sum_{i=1}^{d}z^{t}(i)^{2}}\leq\sqrt{\sum_{i}\operatorname*{\mathbb{E}}z^{t}(i)^{2}}=\sqrt{d\mathrm{Var}(\mathrm{Lap}(b))}=\sqrt{2db^{2}}=2\sqrt{2d}\frac{\lceil\log T\rceil}{\epsilon}~.

Therefore, it suffices to pick

λcϵ22dlogTlogT.\lambda\leq\frac{c~\epsilon}{2\sqrt{2d}\lceil\log T\rceil\log T}~.

For c=αc=\alpha, this is in fact accomplished by the private, accurate market choosing λλ\lambda\leq\lambda^{*} from eq. eq:lambda-star. ∎

Limitations of this result.  Unfortunately, Theorem 2 does not completely solve our problem: the other way that privacy impacts the market’s loss is by lowering the necessary price sensitivity to λ1(logT)2\lambda^{*}\approx\frac{1}{(\log T)^{2}} as mentioned above, leading to a worst-case loss growing with TT. It does not seem possible to address this via a larger transaction fee without giving up incentive to participate: traders participate as long as their expected profit exceeds the fee, and collectively Ω(1/λ)\Omega(1/\lambda) of them can arrive making consistent trades all moving the prices in the same (correct) direction, so the total payout will still be Ω(1/λ)\Omega(1/\lambda).

4 Constant Budget via Adaptive Market Size

In this section, we achieve our original goal by constructing an adaptively-growing prediction market in which each stage, if completed, subsidizes the initial portion of the next.

The market design is the following, with each T(k)T^{(k)} to be chosen later. We run the transaction-fee private market above with T=T(1)T=T^{(1)}, transaction fee α\alpha, and price sensitivity λ(1)=λ(T(1),α/2,γ/2)\lambda^{(1)}=\lambda^{*}(T^{(1)},\alpha/2,\gamma/2) from eq. (1). When (and if) T(1)T^{(1)} participants have arrived, we create a new market whose initial state is such that its prices match the final (noisy) prices of the previous one. We set T(2)T^{(2)} and price sensitivity λ(2)=λ(T(2),α/4,γ/4)\lambda^{(2)}=\lambda^{*}(T^{(2)},\alpha/4,\gamma/4) for the new market. We repeat, halving α\alpha and γ\gamma at each stage and increasing TT in a manner to be specified shortly, until no more participants arrive.

Theorem 3.

For any α,γ,ϵ\alpha,\gamma,\epsilon, the adaptive market satisfies ϵ\epsilon-differential privacy, 2α2\alpha-incentive to trade, (α,γ)(\alpha,\gamma)-accuracy, and a designer budget bound of

BB1722dαϵ(ln4608B12d2γα2ϵ)2,B\leq B_{1}\frac{72\sqrt{2}d}{\alpha~\epsilon}\biggl{(}\ln\frac{4608B_{1}\sqrt{2}d^{2}}{\gamma\alpha^{2}\epsilon}\biggr{)}^{2}~,

where B1B_{1} is the budget bound of the underlying unscaled cost function C1C_{1}.

Proof idea.  We set T(1)=Θ(B1dln(B1d/γαϵ)2α2ϵ)T^{(1)}=\Theta\bigl{(}\frac{B_{1}d\ln(B_{1}d/\gamma\alpha\epsilon)^{2}}{\alpha^{2}~\epsilon}\bigr{)}, and T(k)=4T(k1)T^{(k)}=4T^{(k-1)} thereafter. The key will be the following observation. The total “informational” profit available to the traders (by correcting the initial market prices) is bounded by O(1/λ)O(1/\lambda), so if each trader expects to profit more than the transaction fee cc, then only O(1/λc)O(1/\lambda c) traders can all arrive and simultaneously profit. Indeed, if all TT participants arrive, then the total profit from transaction fees is Θ(T)\Theta(T) while the worst-case loss from the market is O((logT)2)O\left((\log T)^{2}\right).

We can leverage this observation to achieve a bounded worst-case loss with an “adaptive-liquidity” approach, similar in spirit to Abernethy et al. (2014b) but more technically similar to the doubling trick in online learning. Begin by setting λ(1)\lambda^{(1)} on the order of 1/(logT(1))2=Θ(1)1/(\log T^{(1)})^{2}=\Theta(1), and run a private market for T(1)T^{(1)} participants. If fewer than T(1)T^{(1)} participants show up, the worst-case loss is order 1/λ(1)1/\lambda^{(1)}, a constant. If all T(1)T^{(1)} participants arrive, then (for the right choice of constants) the market has actually turned a profit Ω(T(1))\Omega(T^{(1)}) from the transaction fees. Now set up a private market for T(2)=4T(1)T^{(2)}=4T^{(1)} traders with λ(2)\lambda^{(2)} on the order of 1/(logT(2))21/(\log T^{(2)})^{2}. If fewer than T(2)T^{(2)} participants arrive, the worst-case loss is order 1/λ(2)1/\lambda^{(2)}. However, we will have chosen T(2)T^{(2)} such that this loss is smaller than the Ω(T(1))\Omega(T^{(1)}) profit from the previous market. Hence, the total worst-case loss remains bounded by a constant.

If all T(2)T^{(2)} participants arrive, then again this market has turned a profit, which can be used to completely offset the worst-case loss of the next market, and so on. Some complications arise, as to achieve (α,γ)(\alpha,\gamma)-precision, we must set α(1),γ(1),α(2),γ(2),\alpha^{(1)},\gamma^{(1)},\alpha^{(2)},\gamma^{(2)},\dots as a convergent series summing to α\alpha and γ\gamma; and we must show that all of these scalings are possible in such a way that the transaction fees cover the cost of the next iteration. (An interesting direction for future work would be to replace the iterative approach here with the continuous liquidity adaptation of Abernethy et al. (2014b).)

More specifically, we prove that the loss in any round kk that is not completed (not all participants arrive) is at most α16T(k)\frac{\alpha}{16}T^{(k)}; moreover, the profit in any round kk that is completed is at least α2T(k)\frac{\alpha}{2}T^{(k)}. Of course, only one round is not completed: the final round kk. If k=1k=1, then the financial loss is bounded by 1λ(1)\frac{1}{\lambda^{(1)}}, a constant depending only on α,γ,ϵ\alpha,\gamma,\epsilon. Otherwise, the total loss is the sum of the losses across rounds, but the mechanism makes a profit in every round but kk. Moreover, the loss in round kk is at most α2T(k)=α8T(k1)\frac{\alpha}{2}T^{(k)}=\frac{\alpha}{8}T^{(k-1)}, which is at most half of the profit in round k1k-1. So if k2k\geq 2, the mechanism actually turns a net profit.

While this result may seem paradoxical, note that the basic phenomenon appears in a classical (non-private) prediction market with a transaction fee, although to our knowledge this observation has not yet appeared in the literature. Specifically, a classical prediction market with budget bound B1B_{1}, trades of size 11, and a small transaction fee α\alpha, will still have an α\alpha-incentive to participate, and the worst case loss will still be Θ(B1)\Theta(B_{1}); this loss, however, can be extracted by as few as Θ(1)\Theta(1) participants. Any additional participants must be in a sense disagreeing about the correct prices; their transaction fees go toward market maker profit, but they do not contribute further to worst-case loss.

5 Kernels, Buying Data, Online Learning

While preserving privacy in prediction markets is well-motivated in the classical prediction market setting, it is arguably even more important in a setting where machine-learning hypotheses are learned from private personal data. Waggoner et al. Waggoner et al. (2015) develop mechanisms for such a setting based on prediction markets, and further show how to preserve differential privacy of the participants. Yet their mechanisms are not practical in the sense that the financial loss of the mechanism could grow without bound. In this section, we sketch how our bounded-financial-loss market can also be extended to this setting. This yields a mechanism for purchasing data for machine learning that satisfies ϵ\epsilon-differential privacy, α\alpha-precision and incentive to participate, and bounded designer budget.

To develop a mechanism which could be said to “purchase data” from participants, Waggoner et al. Waggoner et al. (2015) extend the classical setting in two ways. The first is to make the market conditional, where we let 𝒵=𝒳×𝒴\mathcal{Z}=\mathcal{X}\times\mathcal{Y}, and have independent markets Cx:dC_{x}:\mathbb{R}^{d}\to\mathbb{R} for each xx. Trades in each market take the form qxdq_{x}\in\mathbb{R}^{d}, which pay out qxϕ(y)q_{x}\cdot\phi(y) upon outcome (x,y)(x^{\prime},y) if x=xx=x^{\prime}, and zero if xxx\neq x^{\prime}. Importantly, upon outcome (x,y)(x,y), only the costs associated to trades in the CxC_{x} market are tallied.

The second is to change the bidding language using a kernel, a positive semidefinite function k:𝒵×𝒵k:\mathcal{Z}\times\mathcal{Z}\to\mathbb{R}. Here we think of contracts as functions f:𝒵f:\mathcal{Z}\to\mathbb{R} in the reproducing kernel Hilbert space (RKHS) \mathcal{F} given by kk, with basis {fz()=k(z,):z𝒵}\{f_{z}(\cdot)=k(z,\cdot)~:~z\in\mathcal{Z}\}. For example, we recover the conditional market setting with independent markets with the kernel k((x,y),(x,y))=𝟏{x=x}ϕ(y)ϕ(y)k((x,y),(x^{\prime},y^{\prime}))=\mathbf{1}\{x=x^{\prime}\}\phi(y)\cdot\phi(y^{\prime}). The RKHS structure is natural here because a basis contract fzf_{z} pays off at each zz^{\prime} according to the “covariance” structure of the kernel, i.e. the payoff of contract fzf_{z} when zz^{\prime} occurs equals fz(z)=k(z,z)f_{z}(z^{\prime})=k(z,z^{\prime}). For example, when 𝒴={1,1}\mathcal{Y}=\{-1,1\} one recovers radial basis classification using k((x,y),(x,y))=yye(xx)2k((x,y),(x^{\prime},y^{\prime}))=yy^{\prime}e^{-(x-x^{\prime})^{2}}.

These two modifications to classical prediction markets, given as Mechanism 2 in Waggoner et al. (2015), have clear advantages as a mechanism to “buy data”. One may imagine that each agent, arriving at time t{1,,T}t\in\{1,\dots,T\}, holds a data point (xt,yt)𝒵=𝒳×𝒴(x^{t},y^{t})\in\mathcal{Z}=\mathcal{X}\times\mathcal{Y}. A natural purchase for this agent would be a basis contract f(xt,yt)f_{(x^{t},y^{t})}, as this corresponds to a payoff that is highest when the test data point actually equals (xt,yt)(x^{t},y^{t}) and decreases with distance as measured by the kernel structure.

The importance of privacy now becomes even more apparent, as the data point (xt,yt)(x^{t},y^{t}) could be information sensitive to trader tt. Fortunately, we can extend our main results to this setting. To demonstrate the idea, we give a sketch of the result and proof below.

Theorem 4 (Informal).

Let 𝒵=𝒳×𝒴\mathcal{Z}=\mathcal{X}\times\mathcal{Y} where 𝒳\mathcal{X} is a compact subset of a finite-dimensional real vector space and 𝒴\mathcal{Y} is finite, and let positive semidefinite kernel k:𝒵×𝒵k:\mathcal{Z}\times\mathcal{Z}\to\mathbb{R} be given. For any choices of accuracy parameters α,γ\alpha,\gamma, privacy parameters ϵ,δ\epsilon,\delta, trade size Δ\Delta, and query limit QQ, the kernel adaptive market satisfies (ϵ,δ)(\epsilon,\delta)-differential privacy, (α,γ)(\alpha,\gamma)-precision, 2α2\alpha-incentive to participate, and a bounded designer budget.

Proof Sketch.

The precision property, i.e. that prices are approximately accurate despite privacy-preserving noise, follows from (Waggoner et al., 2015, Theorem 2), and the technique in Theorem 3 to combine the accuracy and privacy of multiple epochs. The incentive to trade property is essentially unchanged, as a participants’ profit is still the improvement in expected Bregman divergence, which exceeds the transaction fee unless prices are already accurate. It thus remains only to show a bounded designer budget, which is slightly more involved. Briefly, Claim 1 goes through unchanged, and Claim 2 holds as written where now CC becomes CxC_{x} and ztz^{t} becomes zt(x)=ft(x,)z^{t}(x)=f^{t}(x,\cdot), i.e., the trade at time tt restricted to the CxC_{x} market alone.

The remainder of Lemma 1 now proceeds with one modification regarding the constant KK. In eq. (3), the expression for the noise trader loss becomes NTL(λ,T)=𝔼[supx𝒳t=1Tλαtzt(x)2]NTL(\lambda,T^{\prime})=\operatorname*{\mathbb{E}}\bigl{[}\sup_{x\in\mathcal{X}}\sum_{t=1}^{T^{\prime}}\lambda\alpha_{t}\|z^{t}(x)\|_{2}\bigr{]}, where the αt\alpha_{t} are simply coefficients to keep track of how many trades occurred between the buy and sell of noice trade tt. We can proceed as follows:

NTL(λ,T)𝔼[supx1,,xT𝒳t=1Tλαtzt(xt)2]=λt=1Tαt𝔼[supx𝒳zt(x)2]=λt=1TαtK,\displaystyle NTL(\lambda,T^{\prime})\leq\operatorname*{\mathbb{E}}\left[\sup_{x_{1},\ldots,x_{T^{\prime}}\in\mathcal{X}}\sum_{t=1}^{T^{\prime}}\lambda\alpha_{t}\|z^{t}(x_{t})\|_{2}\right]=\lambda\sum_{t=1}^{T^{\prime}}\alpha_{t}\operatorname*{\mathbb{E}}\left[\sup_{x\in\mathcal{X}}\|z^{t}(x)\|_{2}\right]=\lambda\sum_{t=1}^{T^{\prime}}\alpha_{t}K~,

where KK is simply the constant 𝔼[supx𝒳zt(x)2]\operatorname*{\mathbb{E}}\left[\sup_{x\in\mathcal{X}}\|z^{t}(x)\|_{2}\right] where the expectation is taken over the Gaussian process generating the noise. It is well-known that the expected maximum of a Gaussian process is bounded Talagrand (2014), and thus boundedness of KK follows from the fact that 𝒴\mathcal{Y} is finite. Thus, continuing from eq. (3) we obtain NTL(λ,T)TlogT2λKNTL(\lambda,T^{\prime})\leq\frac{T^{\prime}\log T^{\prime}}{2}\lambda K as before, with this new KK. Finally, the proof of Theorem 3 now goes through, as it only treats the mechanism from Theorem 2 as a black box. ∎

We close by noting the similarity between the kernel adaptive market mechanism and traditional learning algorithms, as alluded to in the introduction. As observed by Abernethy, et al. Abernethy et al. (2013), the market price update rule for classical prediction markets resembles Follow-the-Regularized-Leader (FTRL); specifically, the price update at time tt is given by pt=C(qt)=argmaxwΔ(𝒴)w,stdqsR(w)p^{t}=\nabla C(q^{t})=\operatorname*{\text{argmax}}_{w\in\Delta(\mathcal{Y})}\langle w,\sum_{s\leq t}dq^{s}\rangle-R(w), where dqsdq^{s} is the trade at time ss, and R=CR=C^{*} is the convex conjugate of CC.

In our RKHS setting, we can see the same relationship. For concreteness, let Cx(q)=1λC(λq)C_{x}(q)=\tfrac{1}{\lambda}C(\lambda q) for all x𝒳x\in\mathcal{X}, and let R:Δ(𝒴)R:\Delta(\mathcal{Y})\to\mathbb{R} be the conjugate of CC. Suppose further that each agent tt purchases a basis contract dft=fxt,ytdf^{t}=f_{x^{t},y^{t}}, where we take a classification kernel k((x,y),(x,y))=k(x,x)𝟏{y=y}k^{\prime}((x,y),(x^{\prime},y^{\prime}))=k(x,x^{\prime})\mathbf{1}\{y=y^{\prime}\}. Letting dqt(x)=dft(x,)𝒴dq^{t}(x)=df^{t}(x,\cdot)\in\mathbb{R}^{\mathcal{Y}}, the market price at time tt is given by,

pxt\displaystyle p^{t}_{x} =argmaxwΔ(𝒴)w,stdqs(x)1λR(w)\displaystyle=\operatorname*{\text{argmax}}_{w\in\Delta(\mathcal{Y})}\biggl{\langle}w,\sum_{s\leq t}dq^{s}(x)\biggr{\rangle}-\frac{1}{\lambda}R(w)
=argmaxwΔ(𝒴)w,stk((xs,ys),(x,))1λR(w)\displaystyle=\operatorname*{\text{argmax}}_{w\in\Delta(\mathcal{Y})}\biggl{\langle}w,\sum_{s\leq t}k((x^{s},y^{s}),(x,\cdot))\biggr{\rangle}-\frac{1}{\lambda}R(w)
=argmaxwΔ(𝒴)w,stk(xs,x)𝟏ys1λR(w),\displaystyle=\operatorname*{\text{argmax}}_{w\in\Delta(\mathcal{Y})}\biggl{\langle}w,\sum_{s\leq t}k(x^{s},x)\mathbf{1}_{y^{s}}\biggr{\rangle}-\frac{1}{\lambda}R(w)~,

where 𝟏y\mathbf{1}_{y} is an indicator vector. Thus, the market price update follows a natural kernel-weighted FTRL algorithm, where the learning rate λ\lambda is the price sensitivity of the market.

6 Summary and Future Directions

Motivated by the problem of purchasing data, we gave the first bounded-budget prediction market mechanism that achieves privacy, incentive alignment, and precision (low impact of privacy-preserving noise the predictions). To achieve bounded budget, we first introduced and analyzed a transaction fee, achieving a slowly-growing O((logT)2)O((\log T)^{2}) budget bound, thus eliminating the arbitrage opportunities underlying previous impossibility results. Then, observing that this budget still grows in the number of participants TT, we further extended these ideas to design an adaptively-growing market, which does achieve bounded budget along with privacy, incentive, and precision guarantees.

We see several exciting directions for future work. An extension of Theorem 4 where 𝒴\mathcal{Y} need not be finite should be possible via a suitable generalization of Claim 2. Another important direction is to establish privacy for parameterized settings as introduced by Waggoner, et al. Waggoner et al. (2015), where instead of kernels, market participants update the (finite-dimensional) parameters directly as in linear regression. Finally, we would like a deeper understanding of the learning–market connection in nonparametric kernel settings, which could lead to practical improvements for design and deployment.

References

  • Abernethy and Frongillo [2012] J. Abernethy and R. Frongillo. A characterization of scoring rules for linear properties. In Proceedings of the 25th Conference on Learning Theory, pages 1–27, 2012. URL http://jmlr.csail.mit.edu/proceedings/papers/v23/abernethy12/abernethy12.pdf.
  • Abernethy et al. [2013] Jacob Abernethy, Yiling Chen, and Jennifer Wortman Vaughan. Efficient market making via convex optimization, and a connection to online learning. ACM Transactions on Economics and Computation, 1(2):12, 2013. URL http://dl.acm.org/citation.cfm?id=2465777.
  • Abernethy et al. [2014a] Jacob Abernethy, Sindhu Kutty, Sébastien Lahaie, and Rahul Sami. Information aggregation in exponential family markets. In Proceedings of the fifteenth ACM conference on Economics and computation, pages 395–412. ACM, 2014a. URL http://dl.acm.org/citation.cfm?id=2602896.
  • Abernethy and Frongillo [2011] Jacob D. Abernethy and Rafael M. Frongillo. A collaborative mechanism for crowdsourcing prediction problems. In Advances in Neural Information Processing Systems 24, pages 2600–2608, 2011.
  • Abernethy et al. [2014b] Jacob D. Abernethy, Rafael M. Frongillo, Xiaolong Li, and Jennifer Wortman Vaughan. A General Volume-parameterized Market Making Framework. In Proceedings of the Fifteenth ACM Conference on Economics and Computation, EC ’14, pages 413–430, New York, NY, USA, 2014b. ACM. ISBN 978-1-4503-2565-3. doi: 10.1145/2600057.2602900. URL http://doi.acm.org/10.1145/2600057.2602900.
  • Banerjee et al. [2005] A. Banerjee, X. Guo, and H. Wang. On the optimality of conditional expectation as a Bregman predictor. IEEE Transactions on Information Theory, 51(7):2664–2669, July 2005. ISSN 0018-9448. doi: 10.1109/TIT.2005.850145.
  • Chan et al. [2011] T-H Hubert Chan, Elaine Shi, and Dawn Song. Private and continual release of statistics. ACM Transactions on Information and System Security (TISSEC), 14(3):26, 2011.
  • Cummings et al. [2016] Rachel Cummings, David M Pennock, and Jennifer Wortman Vaughan. The possibilities and limitations of private prediction markets. In Proceedings of the 17th ACM Conference on Economics and Computation, EC ’16, pages 143–160. ACM, 2016.
  • Dwork and Roth [2014] Cynthia Dwork and Aaron Roth. The algorithmic foundations of differential privacy. Foundations and Trends in Theoretical Computer Science, 2014.
  • Dwork et al. [2010] Cynthia Dwork, Moni Naor, Toniann Pitassi, and Guy N Rothblum. Differential privacy under continual observation. In Proceedings of the forty-second ACM symposium on Theory of computing, pages 715–724. ACM, 2010.
  • Frongillo et al. [2012] R. Frongillo, N. Della Penna, and M. Reid. Interpreting prediction markets: a stochastic approach. In Advances in Neural Information Processing Systems 25, pages 3275–3283, 2012. URL http://books.nips.cc/papers/files/nips25/NIPS2012_1510.pdf.
  • Frongillo and Reid [2015] Rafael Frongillo and Mark D. Reid. Convergence Analysis of Prediction Markets via Randomized Subspace Descent. In Advances in Neural Information Processing Systems, pages 3016–3024, 2015. URL http://papers.nips.cc/paper/5727-convergence-analysis-of-prediction-markets-via-randomized-subspace-descent.
  • Savage [1971] L.J. Savage. Elicitation of personal probabilities and expectations. Journal of the American Statistical Association, pages 783–801, 1971.
  • Talagrand [2014] Michel Talagrand. Upper and lower bounds for stochastic processes: modern methods and classical problems, volume 60. Springer Science & Business Media, 2014.
  • Waggoner et al. [2015] Bo Waggoner, Rafael Frongillo, and Jacob D Abernethy. A Market Framework for Eliciting Private Data. In Advances in Neural Information Processing Systems 28, pages 3492–3500, 2015. URL http://papers.nips.cc/paper/5995-a-market-framework-for-eliciting-private-data.pdf.

Appendix A Private (Unbounded-Loss) Markets

In this section, we review the private prediction market construction of Waggoner et al. [2015]. We include proofs for completeness and clarity, as we focus on classic, complete cost-function based markets here whereas that paper focused on “kernel markets” which required additional formalism. Our adaptive market will rely on this market construction and results.

Approach and notation.

In a private market, the designer chooses an initial “true” market state q0q^{0} (for convenience, we will assume q0=0q^{0}=0) and announces a “published” market state q^0=0\hat{q}^{0}=0. When participant t=1,,Tt=1,\dots,T arrives and requests trade dqtdq^{t}, the market maker updates the true market state qt=qt1+dqtq^{t}=q^{t-1}+dq^{t}, but does not reveal the true state to anyone. Instead, the market maker announces the published market state q^t\hat{q}^{t}, which is some randomized function of all trades and published market states so far. We assume that dqt1\|dq^{t}\|\leq 1 according to some norm \|\cdot\|, that is, each participant can buy or sell at most one “total” share.444One could also modify our approach to allow arbitrarily large trades, but this would also require adding proportionally large noise in order to continue to preserve privacy. Let \|\cdot\|_{*} denote the dual norm to \|\cdot\|.

Differential privacy.

The market mechanism can be viewed as a randomized function MM that takes as input a list of trades dq=dq1,,dqT\vec{dq}=dq^{1},\dots,dq^{T} and outputs a list of published market states q^0,,q^T\hat{q}^{0},\dots,\hat{q}^{T}. We will call it (ϵ,δ)(\epsilon,\delta)-differentially private if changing a single participant’s trade does not change the distribution on outputs much: if for all dq\vec{dq} and dq\vec{dq^{\prime}} differing only in one entry, and for all (measurable) sets of possible outputs SS,

Pr[M(dq)S]eϵPr[M(dq)S]+δ.\Pr\left[M\left(\vec{dq}\right)\in S\right]\leq e^{\epsilon}\Pr\left[M\left(\vec{dq}^{\prime}\right)\in S\right]~+~\delta.

The mechanism is ϵ\epsilon-differential private if it is (ϵ,0)(\epsilon,0)-d.p. It is reasonable to treat ϵ\epsilon as a constant whose size controls the privacy guarantee, such as ϵ=0.01\epsilon=0.01. Meanwhile, δ\delta is normally preferred to be vanishingly small or 0, as a mechanism can leak the private information of all individuals with δ\delta probability and still be (ϵ,δ)(\epsilon,\delta)-differentially private.

To be careful, we note that the market’s “full output” also includes that it sends each participant their payoff. However, this payoff is a function only of the public noisy market states and of that participant’s trade. The payoff is assumed to be sent privately and separately, unobservable by any other party. By the post-processing property of differential privacy, a trader’s (ϵ,δ)(\epsilon,\delta)-privacy guarantee continues to hold regardless of how the published market states are combined with any side information, even including the full list of all other participant’s trades. (This can be formalized using the notion of joint differential privacy, but for simplicity we will not do so.)

Tool 1: generalized Laplace noise.

Imagine that the market could first collect all TT trades simultaneously, then sum them and publish some q^T\hat{q}^{T}, a noisy version of the market state qT=t=1Tdqtq^{T}=\sum_{t=1}^{T}dq^{t}.

In this scenario, there is only one output q^T\hat{q}^{T} instead of a whole list of outputs q^1,\hat{q}^{1},\dots. The standard, simplest solution to protecting privacy would be to take the true sum qTq^{T} and add noise from a generalization of the Laplace distribution. The real-valued Lap(b)Lap(b) random variable has probability density x12be|x|/bx\mapsto\frac{1}{2b}e^{-|x|/b}. In d\mathbb{R}^{d}, given a norm \|\cdot\|, we define the generalized Lapd(b)Lap_{d}(b) distribution to have probability density proportional to ex/be^{\|x\|/b}. In this case, releasing q^=q+Lapd(1/ϵ)\hat{q}=q+Lap_{d}(1/\epsilon) is ϵ\epsilon-differentially private: Given q,qq,q^{\prime} with qq1\|q-q^{\prime}\|\leq 1, the ratio of probability densities at any q^\hat{q} is eϵqq^/eϵqq^eϵqqeϵe^{\epsilon\|q-\hat{q}\|}/e^{\epsilon\|q^{\prime}-\hat{q}\|}\leq e^{\epsilon\|q-q^{\prime}\|}\leq e^{\epsilon}. (When the norm is L1L^{1}, this corresponds to independent scalar Lap(1/ϵ)Lap(1/\epsilon) noise on each of the dd coordinates.) Note that it also satisfies a good accuracy guarantee, as the amount of noise required does not scale with TT; so with enough participants, this mechanism becomes a very accurate indication of the “average” trade while still preserving privacy.

Tool 2: continual observation technique.

Unfortunately, the above solution is not sufficient because our market must publish a market state at each time step. One naive approach is to apply the above solution independently at each time step, i.e. produce each q^t=qt+zt\hat{q}^{t}=q^{t}+z^{t} where ztz^{t} contains independent Laplace noise. The problem is that each step reveals more information about a trade, for instance, dq1dq^{1} participates in TT separate publications. To continue preserving privacy, each ztz^{t} must have a much larger variance, which makes the published market states very inaccurate.

A second naive approach is to add noise to each dqtdq^{t} just once, producing dq^t=dqt+zt\hat{dq}^{t}=dq^{t}+z^{t}. Then set q^t=s=1tdq^t\hat{q}^{t}=\sum_{s=1}^{t}\hat{dq}^{t}. The benefit to this approach is that it can re-use the noisy ztz^{t} variables across time steps, rather than re-drawing new noise each time. The problem is that, while each ztz^{t} is small in magnitude, there are many of them; for example, the final q^T\hat{q}^{T} contains TT pieces of noise, which add up to a very inaccurate estimate of the true market state. This contrasts with the first naive approach, in which each publication only includes one piece of noise, but that piece of noise is very large.

The idea of the “continual observation” technique, pioneered by Dwork et al. [2010] and Chan et al. [2011], is to strike a balance between these extremes by re-using noise a limited number of times while also keeping each piece of noise small. Roughly, each publication q^t\hat{q}^{t} will include a logarithmic (in tt) number of pieces of noise, each of which is only “logarithmically large”.

Definition 5.

We define the private market mechanism for observation Z𝒵Z\in\mathcal{Z} and securities ϕ:Zd\phi:Z\to\mathbb{R}^{d} using a cost function CC with parameter λ\lambda. At each time tt participant tt arrives and proposes trade dqtdq^{t} with dqt1\|dq^{t}\|\leq 1, unobservable to all others. At most TT participants may arrive. Let qt=j=1tdqjq^{t}=\sum_{j=1}^{t}dq^{j}. Let zt=0z^{t}=0 and for all t1t\geq 1, let ztLapd(2logT/ϵ)z^{t}\sim Lap_{d}(2\lceil\log T\rceil/\epsilon). At each time tt, the mechanism publishes market state

q^t:=qt+zt+zs(t)+zs(s(t))++z0,\hat{q}^{t}:=q^{t}+z^{t}+z^{s(t)}+z^{s(s(t))}+\cdots+z^{0},

where s(t)s(t) is defined by writing the integer tt in binary, then flipping the rightmost “one” bit to zero. Participant tt is charged C(q^t+dqt)C((^q)t)C(\hat{q}^{t}+dq^{t})-C(\hat{(}q)^{t}), unobservable to all others. When outcome ZZ is observed, she is paid dqtϕ(Z)dq^{t}\cdot\phi(Z), unobservable to all others.

We note that s(0)=0s(0)=0 and s(t)<ts(t)<t for all t>0t>0. A convenient notation is to let

q^s(t):t:=(j=s(t)+1tdqj)+zt.\hat{q}^{s(t):t}:=\left(\sum_{j=s(t)+1}^{t}dq^{j}\right)~+~z^{t}.

Then we can define the mechanism recursively as

q^t\displaystyle\hat{q}^{t} =qt+zt+zs(t)+zs(s(t))++z0\displaystyle=q^{t}+z^{t}+z^{s(t)}+z^{s(s(t))}+\cdots+z^{0}
=q^s(t):t+q^s(t).\displaystyle=\hat{q}^{s(t):t}+\hat{q}^{s(t)}.
\bulletdq0dq^{0}\bulletdq1dq^{1}\bulletdq2dq^{2}\bulletdq3dq^{3}\bulletdq4dq^{4}\bulletdq5dq^{5}\bulletdq6dq^{6}\bulletdq7dq^{7}\bulletdq8dq^{8}\bulletdq9dq^{9}\bulletdq10dq^{10}\bulletdq11dq^{11}\bulletdq12dq^{12}\bulletdq13dq^{13}\bulletdq14dq^{14}\bulletdq15dq^{15}\bulletdq16dq^{16}
Figure 1: Picturing the continual observation technique for preserving privacy [Dwork et al., 2010, Chan et al., 2011]. Each dqtdq^{t} is a trade. The true market state at tt is qt=j=1tdqjq^{t}=\sum_{j=1}^{t}dq^{j} and the goal is to release a noisy version q^t\hat{q}^{t} Each arrow originates at tt, points backwards to s(t)s(t), and is labeled with independent Laplace noise vector ztz^{t}. Now q^t=qt+zt+zs(t)+zs(s(t))+\hat{q}^{t}=q^{t}+z^{t}+z^{s(t)}+z^{s(s(t))}+\cdots. In other words, the noise added at tt is a sum of noises obtained by following the arrows all the way back to 0. There are two key properties: Each tt has only logT\log T arrows passing above it, and each path backwards takes only logT\log T jumps.

Remark.

Notice that λ\lambda has no impact on the construction of the market, in particular does not affect the amount of noise to add. Intuitively, this is because the market is defined entirely in “share space”, while price sensitivity relates shares to prices. We will not need to discuss λ\lambda until we discuss accuracy of the prices, which is irrelevant to the proof of privacy.

Theorem 5 (Privacy).

Assuming that all trades satisfy dqt1\|dq^{t}\|\leq 1 (under the same norm as used for the generalized Laplace distribution), the private mechanism is ϵ\epsilon-differentially private in the trades dq1,,dqTdq^{1},\dots,dq^{T} with respect to the output q^1,,q^T\hat{q}^{1},\dots,\hat{q}^{T}.

Proof.

We emphasize that this proof does not contain new ideas beyond the original continual observation technique, but merely adapts them to this setting.

We will imagine that the market publishes every partial sum it uses, i.e. q^s(t):t\hat{q}^{s(t):t} for all time steps tt. The actual published values q^t\hat{q}^{t} are functions of these outputs, so by the post-processing property of differential privacy [Dwork and Roth, 2014], it suffices to show that publishing each of these partial sums would be ϵ\epsilon-differentially private.

The idea is to treat each publication of the form q^s(t):t\hat{q}^{s(t):t} as a separate mechanism, which has (claim 1) a guarantee of (ϵ/logT)(\epsilon/\lceil\log T\rceil)-differential privacy. We then show (claim 2) that any one trade dqtdq^{t} participates in at most logT\lceil\log T\rceil of these mechanisms. These two claims imply the result because, by the composition property of differential privacy [Dwork and Roth, 2014], each trade is therefore guaranteed logT(ϵ/logT)=ϵ\lceil\log T\rceil\cdot(\epsilon/\lceil\log T\rceil)=\epsilon-differential privacy.

First, we claim that each publication q^s(t):t\hat{q}^{s(t):t} preserves ϵ/logT\epsilon/\lceil\log T\rceil differential privacy of each trade dqtdq^{t^{\prime}} that participates, i.e. with s(t)<tts(t)<t^{\prime}\leq t. This follows from the definition of q^s(t):t\hat{q}^{s(t):t} because, since dqt1\|dq^{t^{\prime}}\|\leq 1, an arbitrary change in dqtdq^{t^{\prime}} changes norm of the partial sum of trades by at most 22, and ztz^{t} is a draw from the generalized Lapd(2logT/ϵ)Lap_{d}(2\lceil\log T\rceil/\epsilon) distribution with respect to the same norm.

Second, we claim that each trade dqtdq^{t^{\prime}} participates in at most logT\lceil\log T\rceil different partial sums q^s(t):t\hat{q}^{s(t):t}. To show this, we only need to count the time steps tt where s(t)<tts(t)<t^{\prime}\leq t, in other words, integers ttt\geq t^{\prime} where zeroing the rightmost “one” bit gives a number less than tt^{\prime}.

Without loss of generality, the the binary expansion of tt is bmbm1bj100b_{m}b_{m-1}\ldots b_{j}10\ldots 0 for some m,jm,j and then s(t)s(t) has expansion bmbm1bj000b_{m}b_{m-1}\ldots b_{j}00\ldots 0. Hence the condition s(t)<tts(t)<t^{\prime}\leq t implies that the binary expansion of tt matches that of tt^{\prime} from bits mm to jj, then has a one at bit j1j-1, and has zeroes at all lower-order bits. Since mm is fixed for tt^{\prime}, this can only happen once for each jj, or at most mm total times; and mlogTm\leq\lceil\log T\rceil because tTt^{\prime}\leq T. ∎

Lemma 2 (Accuracy of share vector).

In the private mechanism with L1L^{1} norm, dd securities, and TT time steps, we have with probability 1γ1-\gamma,

maxtqtq^t142dlogTϵln(2Tdγ).\max_{t}\|q^{t}-\hat{q}^{t}\|_{1}\leq\frac{4\sqrt{2}d\log\lceil T\rceil}{\epsilon}\ln\left(\frac{2Td}{\gamma}\right).
Proof.

For each tt, each coordinate ii of qtq^tq^{t}-\hat{q}^{t} is the sum of at most logT\lceil\log T\rceil independent variables distributed Lap(2logT/ϵ)Lap(2\lceil\log T\rceil/\epsilon). We will choose β\beta such that each coordinate’s absolute value exceeds β\beta with probability γTd\frac{\gamma}{Td}; there are dd coordinates per time step and TT time steps, so a union bound gives the result.

Choose β\beta such that, if YY is the sum of k=logTk=\lceil\log T\rceil independent Lap(b)Lap(b) variables with b=2logT/ϵ)b=2\lceil\log T\rceil/\epsilon) variables, then

Pr[|Y|>β]=γ.\Pr[|Y|>\beta]=\gamma^{\prime}.

A concentration bound for the sum of kk independent Lap(b)Lap(b) variables, Corollary 12.3 of Dwork and Roth [2014]555In the parameters of that Corollary, we choose ν=bln(2/γ)\nu=b\sqrt{\ln(2/\gamma^{\prime})} as we will have ln(2/γ)>k\ln(2/\gamma^{\prime})>k. gives

β22bln2γ.\beta\leq 2\sqrt{2}b\ln\frac{2}{\gamma^{\prime}}.

Now choose γ=γTd\gamma^{\prime}=\frac{\gamma}{Td}. To recap, each |qt(i)q^t(i)|β|q^{t}(i)-\hat{q}^{t}(i)|\leq\beta except with probability γ=γTd\gamma^{\prime}=\frac{\gamma}{Td}, hence by a union bound this holds for all t,it,i except with probability γ\gamma, hence qtq^t1dβ\|q^{t}-\hat{q}^{t}\|_{1}\leq d\beta except with probability γ\gamma. ∎

As mentioned above, the previous results (Theorem 5 and Lemma 2) do not depend on λ\lambda at all, because they do not mention the prices. We now ask what a “reasonable” choice of λ\lambda can be so that the prices are interpretable as predictions, i.e. the prices are “accurate”.

Theorem 6 (Accuracy of prices).

In the private mechanism, let pt=C(qt)p^{t}=\nabla C(q^{t}) and let p^t=C(q^t)\hat{p}^{t}=\nabla C(\hat{q}^{t}). Then to satisfy ptp^t1α\|p^{t}-\hat{p}^{t}\|_{1}\leq\alpha for all tt, except with probability γ\gamma, it suffices for the price sensitivity to be

λ=αϵ42dlogTln(2Td/γ).\lambda^{*}=\frac{\alpha~\epsilon}{4\sqrt{2}d\lceil\log T\rceil\ln(2Td/\gamma)}~.
Proof.

By definition of λ\lambda, we have

ptp^t1\displaystyle\|p^{t}-\hat{p}^{t}\|_{1} λqtq^t1\displaystyle\leq\lambda\|q^{t}-\hat{q}^{t}\|_{1}
λ42dlogTϵln(2Tdγ)\displaystyle\leq\lambda\frac{4\sqrt{2}d\lceil\log T\rceil}{\epsilon}\ln\left(\frac{2Td}{\gamma}\right) (5)

for all tt except with probability γ\gamma, by Lemma 2. We now just choose λ\lambda so that (5)α(\ref{eqn:lambda-for-accuracy})\leq\alpha. ∎

Appendix B Slowly-Growing Budget

We prove the incentive and budget claims in separate lemmas.

Lemma 3 (Incentive to trade).

In the private market with transaction fee α\alpha, a participant at time tt having belief pp with pp^t2α\|p-\hat{p}^{t}\|_{\infty}\geq 2\alpha can make a strictly positive expected profit by participating.

Proof.

Ignoring the transaction fee, the expected profit from purchasing dqdq (recall dq11\|dq\|_{1}\leq 1) is

profit =dq,p+C(q^t)C(q^t+dq).\displaystyle=\langle dq,p\rangle+C(\hat{q}^{t})-C(\hat{q}^{t}+dq).

Because CC is convex, C(q^t)C(q^t+dq)C(q^t+dq),dqC(\hat{q}^{t})-C(\hat{q}^{t}+dq)\geq\langle\nabla C(\hat{q}^{t}+dq),-dq\rangle. So

profit dq,pC(q^t+dq)\displaystyle\geq\langle dq,p-\nabla C(\hat{q}^{t}+dq)\rangle
=dq,pp^tdq,C(q^t+dq)p^t.\displaystyle=\langle dq,p-\hat{p}^{t}\rangle-\langle dq,\nabla C(\hat{q}^{t}+dq)-\hat{p}^{t}\rangle.

By Hölder’s inequality and the definition of price sensitivity,

dq,C(q^t+dq)p^t\displaystyle\langle dq,\nabla C(\hat{q}^{t}+dq)-\hat{p}^{t}\rangle dqC(q^+dq)p^1\displaystyle\leq\|dq\|_{\infty}\|\nabla C(\hat{q}+dq)-\hat{p}\|_{1}
dqλdq1\displaystyle\leq\|dq\|_{\infty}\lambda\|dq\|_{1}
λ.\displaystyle\leq\lambda.

So we have

max profit maxdq:dq11dq,pp^tλ\displaystyle\geq\max_{dq:\|dq\|_{1}\leq 1}\langle dq,p-\hat{p}^{t}\rangle-\lambda
=pp^λ\displaystyle=\|p-\hat{p}\|_{\infty}-\lambda
2αλ\displaystyle\geq 2\alpha-\lambda
>α\displaystyle>\alpha

as λ<α\lambda<\alpha by construction. Because there exists a trade with expected profit strictly above α\alpha, the trader has an incentive to pay the α\alpha transaction fee and participate. ∎

Claim 3 (Claim 1).

For each tt, exactly b(t)b(t) traders arrive between the purchase and the sale of bundle ztz^{t}; furthermore, qselltqbuytq_{\text{sell}}^{t}-q_{\text{buy}}^{t} is exactly equal to the sum of these participants’ trades.

Proof.

Note that if we write tt in binary, it has a one in the bit position logb(t)\log b(t), followed by zeros. By definition of the algorithm, ztz^{t} is sold at the next time t>tt^{\prime}>t where the bit logb(t)\log b(t) is flipped to zero. So we have tt=b(t)t^{\prime}-t=b(t), so b(t)b(t) traders have arrived.

Now we want to show that qselltqbuytq_{\text{sell}}^{t}-q_{\text{buy}}^{t} is the sum of their trades, i.e. that every noise trade bundle ztz^{t^{\prime}} held by the trader before buying ztz^{t} is still held at the moment of selling ztz^{t}, and no other noise trade bundles are held at that time.

Consider all of the noise bundles that were already held at time tt (after selling the appropriate bundles at that time, but before purchasing ztz^{t}). By definition of the algorithm, these were purchased at times ss with b(s)>b(t)b(s)>b(t), so by the above discussion, they are not sold until times tt^{\prime} where the bits in positions logb(s)\log b(s) are flipped to zero, which cannot happen until after bit logb(t)\log b(t) is flipped and ztz^{t} is sold. Meanwhile, every bundle purchased after ztz^{t} is sold by, at the latest, the same time that ztz^{t} is sold, as they correspond to lower-order bits; and any sold at the same time as ztz^{t} are sold first because they were purchased later. ∎

Claim 4 (Claim 2).

If the noise trader purchases and later sells ztz^{t}, then her net loss in expectation over ztz^{t} (but for any trader behavior in response to ztz^{t}), is at most λb(t)K\lambda b(t)K where K=𝔼zt2K=\operatorname*{\mathbb{E}}\|z^{t}\|_{2}.

Proof.

Given the noise trader’s bundle drawn is ztz^{t}, her loss is:

C(qbuyt+zt)C(qbuyt)+C(qsellt)C(qsellt+zt).C(q_{\text{buy}}^{t}+z^{t})-C(q_{\text{buy}}^{t})+C(q_{\text{sell}}^{t})-C(q_{\text{sell}}^{t}+z^{t}).

The first pair of terms represents the payment made to purchase ztz^{t} (moving the market state from qbuytq_{\text{buy}}^{t}); the second pair represents the payment to sell ztz^{t} (moving the state to qselltq_{\text{sell}}^{t}). Claim 3 implies that qbuytqsellt1b(t)\|q_{\text{buy}}^{t}-q_{\text{sell}}^{t}\|_{1}\leq b(t), as each trader can buy or sell at most 11 unit of shares. Therefore, the net loss on bundle ztz^{t} is at most

𝔼ztmaxq,q:qq1b(t)C(q+zt)C(q)+C(q)C(q+zt)\displaystyle\operatorname*{\mathbb{E}}_{z^{t}}~\max_{q,q^{\prime}:\|q-q^{\prime}\|_{1}\leq b(t)}C(q+z^{t})-C(q)+C(q^{\prime})-C(q^{\prime}+z^{t})

Now, we have

C(q+zt)C(q)\displaystyle C(q+z^{t})-C(q) =x=01C(q+xzt)zt𝑑x,\displaystyle=\int_{x=0}^{1}\nabla C(q+xz^{t})\cdot z^{t}dx,
C(q+r+zt)C(q+r)\displaystyle C(q+r+z^{t})-C(q+r) =x=01C(q+r+xzt)zt𝑑x.\displaystyle=\int_{x=0}^{1}\nabla C(q+r+xz^{t})\cdot z^{t}dx.

So the difference is

x=01C(q+xzt)C(q+r+xzt),zt𝑑x\displaystyle\int_{x=0}^{1}\left\langle\nabla C(q+xz^{t})-\nabla C(q+r+xz^{t})~,~z^{t}\right\rangle dx
x=01λr2zt2𝑑x\displaystyle\leq\int_{x=0}^{1}\lambda\|r\|_{2}\|z^{t}\|_{2}dx
=λr2zt2\displaystyle=\lambda\|r\|_{2}\|z^{t}\|_{2}

by definition of price sensitivity λ\lambda. We also have r2r1b(t)\|r\|_{2}\leq\|r\|_{1}\leq b(t). This bound holds for each outcome of ztz^{t} and any behavior of the participants, so we conclude the lemma statement, that expected loss is bounded by λb(t)𝔼zt2\lambda b(t)\operatorname*{\mathbb{E}}\|z^{t}\|_{2}. ∎

Appendix C Constant Budget Bound

Lemma 4.

Let A,DA,D be constants at least 11 with AD5AD\geq 5. Then for all T9A(ln(AD))2T\geq 9A\left(\ln(AD)\right)^{2}, we have TA(ln(TD))2T\geq A\left(\ln(TD)\right)^{2}.

Proof.

Let T=9A(ln(AD))2T^{*}=9A\left(\ln(AD)\right)^{2}. First, we prove the inequality for TT^{*}:

T\displaystyle T^{*} =9A(ln(AD))2\displaystyle=9A\left(\ln(AD)\right)^{2}
=A(3ln(AD))2\displaystyle=A\left(3\ln(AD)\right)^{2}
=A(ln(AD(AD)2))2.\displaystyle=A\left(\ln(AD(AD)^{2})\right)^{2}.

Now for all AD5AD\geq 5, we have AD3ln(AD)AD\geq 3\ln(AD), so

T\displaystyle T^{*} A(ln(9AD(ln(AD))2))2\displaystyle\geq A\left(\ln\left(9AD(\ln(AD))^{2}\right)\right)^{2}
=A(ln(TD))2,\displaystyle=A\left(\ln(T^{*}D)\right)^{2},

as desired. Now we wish to extend this to all TTT\geq T^{*}. Compare the derivative of the left side, dTdT=1\frac{dT}{dT}=1, with that of the right side:

ddT(A(ln(TD))2)\displaystyle\frac{d}{dT}\left(A\left(\ln(TD)\right)^{2}\right) =2Aln(TD)T\displaystyle=\frac{2A\ln(TD)}{T}
2ln(TD)1\displaystyle\leq\frac{2}{\ln(TD)}\leq 1

at T=TT=T^{*}. Now if the inequality holds for all T[T,T)T^{\prime}\in[T^{*},T), then it holds for TT as the left side only increases more quickly than the right. So by transfinite induction, it holds for all TTT\geq T^{*}. ∎

Proof of Theorem 3.

Fix the parameters ϵ\epsilon and dd throughout. Let λ(T,α,γ)\lambda^{*}(T,\alpha,\gamma) be the price sensitivity parameter as a function of these variables given in Theorem 6.

ϵ\epsilon-differential privacy of the market follows by the post-processing property of differential privacy [Dwork and Roth, 2014] because each stage kk is differentially private for the participants who arrive in that stage. All information released after stage kk depends on these participants only through the noisy market state at the end of stage kk, which is ϵ\epsilon-d.p.

To show the incentive guarantee, note that the transaction fee is always fixed at α\alpha, so the incentive proof of Theorem 2 goes through immediately.

To show the accuracy guarantee, note the the prices up to T(1)T^{(1)} arrivals satisfy an α/2\alpha/2 guarantee; therefore the starting prices of the new market are within α/2\alpha/2 of what they would be without added noise. The prices up to T(2)T^{(2)} additional arrivals are within α/2+α/4\alpha/2+\alpha/4 of what they would have been (since they begin within α/2\alpha/2 and are designed to stay within α/4\alpha/4 of this shifted goal); and so on, telescoping to at most α\alpha. Similarly, the chance of failure of any of these guarantees, by a union bound, is at most γ/2+γ/4+γ\gamma/2+\gamma/4+\cdots\leq\gamma.

Now we must show bounded worst-case loss, and how to set T(k)T^{(k)}. We will choose T(1)T^{(1)} to be a constant and each T(k)=4T(k1)T^{(k)}=4T^{(k-1)}.

We will claim two things:

  1. 1.

    In the final stage kk where not all participants arrive, the market maker’s loss is at most α16T(k)\frac{\alpha}{16}T^{(k)}.

  2. 2.

    In each stage kk that is completed (all T(k)T^{(k)} participants arrive), the market maker’s profit from that stage is at least α2T(k)\frac{\alpha}{2}T^{(k)}.

These together prove bounded worst-case loss: If at least one stage is completed, the total profit is in fact positive: it is positive from all but the last stage, whose loss is at most α16T(k)α4T(k1)\frac{\alpha}{16}T^{(k)}\leq\frac{\alpha}{4}T^{(k-1)} which is smaller than the profit made in stage k1k-1. If no stages are completed, i.e. fewer than T(1)T^{(1)} participants arrive, then expected worst-case loss is bounded by Bλ(1)\frac{B^{\prime}}{\lambda^{(1)}}. This gives a budget bound of Bλ(1)\frac{B^{\prime}}{\lambda^{(1)}}, which will be computed below.

Proof of (1).

First, we must prove that the worst-case loss in stage 11 is at most α16T(1)\frac{\alpha}{16}T^{(1)}. In doing so, we will explicitly compute a sufficient T(1)T^{(1)} and this worst-case loss. Then, we must show the same fact for all other stages.

The worst-case loss in stage 11, by Theorem 2, is

B\displaystyle B =Bλ(1)\displaystyle=\frac{B^{\prime}}{\lambda^{(1)}}
=B82dlogT(1)ln(4T(1)d/γ)αϵ\displaystyle=B^{\prime}\frac{8\sqrt{2}d\lceil\log T^{(1)}\rceil\ln\left(4T^{(1)}d/\gamma\right)}{\alpha~\epsilon}
B82d(ln(4T(1)d/γ))2αϵ.\displaystyle\leq B^{\prime}\frac{8\sqrt{2}d\left(\ln\left(4T^{(1)}d/\gamma\right)\right)^{2}}{\alpha~\epsilon}.

For convenience, set A=B82dαϵA^{\prime}=B^{\prime}\frac{8\sqrt{2}d}{\alpha\epsilon} and set D=4d/γD=4d/\gamma. Then we have BA(ln(T(1)D))2B\leq A^{\prime}\left(\ln(T^{(1)}D)\right)^{2}; to prove claim (1) for stage 11, we wish to pick T(1)T^{(1)} such that Bα16T(1)B\leq\frac{\alpha}{16}T^{(1)}. Setting A=16αAA=\frac{16}{\alpha}A^{\prime}, we need to have A(ln(T(1)D))2T(1)A\left(\ln(T^{(1)}D)\right)^{2}\leq T^{(1)}. By Lemma 4, this holds for

T(1)\displaystyle T^{(1)} =9A(ln(AD))2\displaystyle=9A\left(\ln(AD)\right)^{2}
=B11522d(ln4608B2d2γα2ϵ)2α2ϵ\displaystyle=B^{\prime}\frac{1152\sqrt{2}d\left(\ln\frac{4608B^{\prime}\sqrt{2}d^{2}}{\gamma\alpha^{2}\epsilon}\right)^{2}}{\alpha^{2}~\epsilon}

So for this choice of T(1)T^{(1)}, we have claim (1) for stage 11. We note the budget bound is Bα16T(1)B\leq\frac{\alpha}{16}T^{(1)}. Now we just show that T(k)T^{(k)} increases faster than 1/λ(k)1/\lambda^{(k)}. T(k)=4T(k1)T^{(k)}=4T^{(k-1)}, but

Bλ(k)\displaystyle\frac{B}{\lambda^{(k)}} =42B2kdlogT(k)ln(2T(k)d2k/γ)αϵ\displaystyle=\frac{4\sqrt{2}B2^{k}d\lceil\log T^{(k)}\rceil\ln\left(2T^{(k)}d2^{k}/\gamma\right)}{\alpha~\epsilon}
=242B2k1d2+logT(k1)(ln(8)+ln(2T(k1)d2k1/γ))αϵ\displaystyle=2\frac{4\sqrt{2}B2^{k-1}d\lceil 2+\log T^{(k-1)}\rceil\left(\ln(8)+\ln\left(2T^{(k-1)}d2^{k-1}/\gamma\right)\right)}{\alpha~\epsilon}
41λ(k1)\displaystyle\leq 4\frac{1}{\lambda^{(k-1)}}

for sufficiently large T(k1)T^{(k-1)}, i.e. if T(1)T^{(1)} is a sufficiently large constant. So Bλ(k)\frac{B}{\lambda^{(k)}} grows more slowly than T(k)T^{(k)} and the inequality Bλ(k)α16T(k)\frac{B}{\lambda^{(k)}}\leq\frac{\alpha}{16}T^{(k)} continues to hold.

Proof of (2).

Let us lower-bound the profit in stage kk if completed. By Inequality 4 (from Theorem 2), that the market-maker profit if T=T(k)T^{\prime}=T^{(k)} participants arrive is

T(k)(cKλ(k)logT(k))Bλ(k)\displaystyle T^{(k)}\left(c-K\lambda^{(k)}\log T^{(k)}\right)-\frac{B}{\lambda^{(k)}}
=T(k)(α2dlogT(k)logT(k)ϵ(α/2k)ϵ42dlogT(k)ln(2dT(k)2k/γ))Bλ(k)\displaystyle=T^{(k)}\left(\alpha-\frac{\sqrt{2d}\lceil\log T^{(k)}\rceil\log T^{(k)}}{\epsilon}\frac{(\alpha/2^{k})~\epsilon}{4\sqrt{2}d\lceil\log T^{(k)}\rceil\ln(2dT^{(k)}2^{k}/\gamma)}\right)-\frac{B}{\lambda^{(k)}}
=αT(k)(1logT(k)4(2k)dln(2dT(k)2k/γ))Bλ(k).\displaystyle=\alpha T^{(k)}\left(1-\frac{\log T^{(k)}}{4(2^{k})\sqrt{d}\ln(2dT^{(k)}2^{k}/\gamma)}\right)-\frac{B}{\lambda^{(k)}}.

Recall that Bλ(k)α16T(k)\frac{B}{\lambda^{(k)}}\leq\frac{\alpha}{16}T^{(k)}. We want to conclude that the profit in stage kk is at least α2T(k)\frac{\alpha}{2}T^{(k)}, so we just need to show that

1logT(k)4(2k)dln(2dT(k)2k/γ)11612.1~-~\frac{\log T^{(k)}}{4(2^{k})\sqrt{d}\ln(2dT^{(k)}2^{k}/\gamma)}~-\frac{1}{16}\geq\frac{1}{2}.

The fraction is decreasing in kk, so it suffices to achieve this for k=1k=1, d=1d=1, and γ=1\gamma=1, where we have

logT(1)8ln(4T(1))14.\frac{\log T^{(1)}}{8\ln(4T^{(1)})}\leq\frac{1}{4}.

This suffices to prove Claim (2). ∎