This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11footnotetext: osogami@jp.ibm.com22footnotetext: hirotak@g.ecc.u-tokyo.ac.jp; This research was conducted at IBM Research.

Mechanism design with multi-armed bandit

Takayuki Osogami IBM Research - Tokyo Hirota Kinoshita The University of Tokyo Segev Wasserkrug IBM Research - Israel
Abstract

A popular approach of automated mechanism design is to formulate a linear program (LP) whose solution gives a mechanism with desired properties. We analytically derive a class of optimal solutions for such an LP that gives mechanisms achieving standard properties of efficiency, incentive compatibility, strong budget balance (SBB), and individual rationality (IR), where SBB and IR are satisfied in expectation. Notably, our solutions are represented by an exponentially smaller number of essential variables than the original variables of LP. Our solutions, however, involve a term whose exact evaluation requires solving a certain optimization problem exponentially many times as the number of players, NN, grows. We thus evaluate this term by modeling it as the problem of estimating the mean reward of the best arm in multi-armed bandit (MAB), propose a Probably and Approximately Correct estimator, and prove its asymptotic optimality by establishing a lower bound on its sample complexity. This MAB approach reduces the number of times the optimization problem is solved from exponential to O(NlogN)O(N\,\log N). Numerical experiments show that the proposed approach finds mechanisms that are guaranteed to achieve desired properties with high probability for environments with up to 128 players, which substantially improves upon the prior work.

1 Introduction

Multi-agent systems can be made efficient by mediators who make system-wide (social) decisions in a way that maximizes social welfare. For example, in a trading network where firms sell goods to each other or to external markets (Hatfield et al.,, 2013), the mediator can ensure that those traded goods are produced by the firms with the lowest costs and purchased by those with the highest needs (Osogami et al.,, 2023). While such mediators could maximize their own profit by charging participants (e.g., today’s tech giants who operate consumer marketplaces or ads platforms), they would typically take most of the profit, leaving the participants with only a small portion.

We instead envision an open platform whose purpose is to provide maximal benefits to the participants in multi-agent systems. This is similar to the purpose of designing an auction mechanism for public resources, which should be given to those who need them most, and there should be neither budget deficits nor surpluses on the mediator (Bailey,, 1997; Cavallo,, 2006; Dufton et al.,, 2021; Gujar and Narahari,, 2011; Guo,, 2012; Guo and Conitzer,, 2009; Manisha et al.,, 2018; Tacchetti et al.,, 2022). However, such mechanisms exploit the particular structure of single-sided auctions, where all participants are buyers, and it may be impossible to achieve desired properties in other multi-agent systems such as double-sided auctions (Hobbs et al.,, 2000; Zou,, 2009; Widmer and Leukel,, 2016; Stößer et al.,, 2010; Kumar et al.,, 2018; Chichin et al.,, 2017), matching markets (Zhang and Zhu,, 2020), and trading networks (Osogami et al.,, 2023; Wasserkrug and Osogami,, 2023) even if those properties could be guaranteed for single-sided auctions.

Here, we design mechanisms for general environments that include all these multi-agent systems, with the primary objectives of efficiency and strong budget balance (SBB). Specifically, we require that the mediator chooses a social decision such that its total value to the players (or participants) is maximized (decision efficiency; DE) and that the expected revenue of the mediator is equal to a target value, ρ\rho\in\mathbb{R} (SBB when ρ=0\rho=0). As is standard in Bayesian mechanism design (Shoham and Leyton-Brown,, 2009), we assume that the value of each social decision to a player depends on the player’s type that is only known to that player, but a joint probability distribution of the players’ types is a common knowledge. Since it would be difficult to achieve DE without knowing the types, we also require that the optimal strategy of each player is to truthfully declare its type regardless of the strategies of the others (dominant strategy incentive compatibility; DSIC). To promote participation, we also require that the expected utility of each player is no less than a target value, θ(tn)\theta(t_{n})\in\mathbb{R}, that can depend on the type tnt_{n} of the player (individual rationality or IR when θ0\theta\equiv 0).

Although these properties are standard in the literature of mechanism design (Shoham and Leyton-Brown,, 2009), here we have introduced parameters, ρ\rho and θ\theta, to generalize the standard definitions of SBB (ρ=0\rho=0) and IR (θ0\theta\equiv 0) with three motivations. First, it is not always possible to achieve the four desired properties with the standard definitions of IR and SBB (Green and Laffont,, 1977; Myerson and Satterthwaite,, 1983; Osogami et al.,, 2023), while the generalization will enable the exact characterization of when those properties can be satisfied. Second, this generalization will allow us to develop a principled approach of learning a mechanism, where some of the quantities are estimated from samples, with theoretical guarantees. The mechanism designed with our learning approach can be shown to satisfy the four desired properties with high probability. Finally, the additional parameters will allow us to model practical requirements. For example, the mediator might need positive revenue to cover the cost of maintaining the platform or might want to guarantee positive expected utility to players to encourage participation in this platform rather than others.

We require SBB and IR in expectation (ex ante or interim) with respect to the distribution of types, while DE and DSIC are satisfied for any realization of types (ex post). While these assumptions are similar to those in Osogami et al., (2023); Wasserkrug and Osogami, (2023) for trading networks, the in expectation properties are certainly weaker than the ex post properties usually assumed for auctions (Bailey,, 1997; Cavallo,, 2006; Dufton et al.,, 2021; Gujar and Narahari,, 2011; Guo,, 2012; Guo and Conitzer,, 2009; Manisha et al.,, 2018; Tacchetti et al.,, 2022). With the weaker properties, however, we derive analytical solutions of the mechanisms that satisfy all the four desired properties in the general environments (and characterize when those properties can be satisfied). This is in stark contrast to the prior work, where mechanisms are analytically derived only for auctions with a single type of goods (Bailey,, 1997; Cavallo,, 2006; Guo,, 2011; Guo and Conitzer,, 2007, 2009; Moulin,, 2009) or with unit demand (Gujar and Narahari,, 2011; Guo,, 2012). For more complex auctions (Dufton et al.,, 2021; Manisha et al.,, 2018; Tacchetti et al.,, 2022) or trading networks (Osogami et al.,, 2023; Wasserkrug and Osogami,, 2023), mechanisms are computed by numerically solving optimization problems, whose size often grows exponentially with the number of players.

While the numerical approaches proposed in Osogami et al., (2023) have been applied only to the trading networks with two players, our analytical solutions can be evaluated numerically with 10\sim 10 players, depending on the number of types. The key bottleneck in our analytical solutions lies in the evaluation of the minimum expected value over possible types of each player. Exact evaluation of this quantity would require computing an efficient social decision for all the KNK^{N} combinations of types, where KK is the number of possible types of each player, and NN is the number of players.

To overcome this bottleneck, we model the problem of evaluating this minimum expected value in a multi-armed bandit (MAB) approach (Lattimore and Szepesvári,, 2020) and propose an asymptotically optimal learning algorithm for this problem. While the standard objectives of MAB are regret minimization (Auer et al.,, 1995, 2002) and best arm identification (Audibert et al.,, 2010; Maron and Moore,, 1993; Mnih et al.,, 2008; Bubeck et al.,, 2009), our objective is to estimate the mean reward of the best arm. We propose a probably approximately correct (PAC) algorithm, which approximately (with error at most ε\varepsilon) estimates the best mean with high probability (at least 1δ1-\delta), and proves that its sample complexity, O((K/ε2)log(1/δ))O((K/\varepsilon^{2})\,\log(1/\delta)), matches the lower bound that we derive. This learning approach substantially reduces the number of computing efficient social decisions from KNK^{N} to O(KNlogN)O(K\,N\,\log N), enabling us to numerically find mechanisms for 100\sim 100 players, depending on KK.

Our contributions thus revolve around the optimization problem whose solution gives the mechanism that satisfies DE, SBB, DSIC, and IR (see Section 3-4). First, we establish a sufficient condition that ensures the optimization problem has feasible solutions, and prove that this sufficient condition is also necessary when the players have independent types (see Section 5). Second, for cases where this sufficient condition holds, we analytically derive a class of optimal solutions to this optimization problem, which in turn gives mechanisms that satisfy DE, SBB, DSIC, and IR for general environments including auctions and trading networks (see Section 5). Third, we model the problem of evaluating a quantity in the above analytical expressions as best mean estimation in MAB, propose a PAC algorithm for this problem, and prove its asymptotic optimality (see Section 6). Finally, we empirically validate the effectiveness of the proposed approach (see Section 7). In Section 2, we start by positioning our contributions to the prior work.

2 Related work

The prior work most related to ours is Osogami et al., (2023), which formulates and numerically solves the LP whose solution gives the mechanism for a trading network that satisfies DE, DSIC, IR, and weak budget balance (WBB; expected revenue of the mediator is nonnegative). While the objective of the LP is rather arbitrary and SBB is given just as an example in Osogami et al., (2023), we focus on SBB and drive analytical solutions for this particular objective. Our formulation extends Osogami et al., (2023) with additional parameters and notations that cover environments beyond trading networks, but these extensions are relatively straightforward.

In the rest of this section, we discuss related work on mechanism design, with a focus on those aiming to achieve SBB, as well as related work on MAB with a focus on PAC algorithms. In particular, we will see that our learning approach is unique in that we estimate a particular quantity in the optimal solution that we derive analytically, and this leads us to propose a new PAC algorithm and establish its optimality for an underexplored objective of best mean estimation in MAB.

In single-sided auctions where only buyers make strategic decisions, Vickrey–Clarke–Groves (VCG) mechanisms with Clark’s pivot rule (also called VCG auction) satisfy ex post DE (called allocative efficiency in auctions), DSIC, IR, and WBB (Nisan,, 2007). However, the Green-Laffont Impossibility Theorem implies that no mechanism can guarantee DE, DSIC, and SBB simultaneously for all environments (Green and Laffont,, 1977, 1979). This has led to a series of work on redistributing the revenue of the mediator to the players as much as possible (i.e., to make budget balance as strong as possible), while satisfying DSIC, DE, IR, and WBB. For auctions with single or homogeneous goods (Bailey,, 1997; Cavallo,, 2006; Guo,, 2011; Guo and Conitzer,, 2007, 2009; Moulin,, 2009) or for auctions where players have unit demand (Gujar and Narahari,, 2011; Guo,, 2012), researchers have derived analytical solutions that optimally redistribute the payment to the players. For auctions with multi-unit demands on heterogeneous goods, the prior work has proposed numerical approaches that seek to find the piecewise linear functions (Dufton et al.,, 2021) or neural networks (Manisha et al.,, 2018; Tacchetti et al.,, 2022) that best approximate the optimal redistribution functions.

We consider the environments that not only allow heterogeneous goods and multi-unit demands but also are more general than single-sided auctions. In particular, our players may have negative valuation on a social decision. The Myerson-Satterthwaite Impossibility Theorem (Myerson and Satterthwaite,, 1983) thus implies that, unlike VCG auctions, no mechanism can guarantee ex post DE, DSIC, IR, and WBB simultaneously for all the environments that we consider. We thus derive mechanisms that achieve DE, DSIC, IR, and SBB in the best possible manner. A limitation of our results is that IR and SBB are satisfied only in expectation. Such a guarantee in expectation can however be justified for risk-neutral mediator and players (Osogami et al.,, 2023). Our model can also guarantee strictly positive expected utility, which in turn can ensure nonnegative utility with high probability when the player repeatedly participate in the mechanism.

For auctions, there also exists a large body of the literature on maximizing the revenue of the mediator (Myerson,, 1981) with recent focus on automated mechanism design (AMD) with machine learning (Duetting et al.,, 2019; Rahme et al.,, 2021; Ivanov et al.,, 2022; Curry et al.,, 2020) and analysis of its sample complexity (Balcan et al.,, 2016; Morgenstern and Roughgarden,, 2015; Syrgkanis,, 2017). Similar to these and other studies of AMD (Sandholm,, 2003; Conitzer and Sandholm,, 2002), we formulate an optimization problem whose solution gives the mechanism with desired properties. However, instead of solving it numerically, we analytically derive optimal solutions. Also, while the prior work analyzes the sample complexity for finding the mechanism that maximizes the expected revenue, we analytically find optimal mechanisms and characterize the sample complexity of evaluating a particular expression in the analytically designed mechanisms.

We evaluate our analytical expression through best mean estimation (BME) in MAB, where the standard objectives are regret minimization (Auer et al.,, 1995, 2002) and best arm identification (BAI) (Audibert et al.,, 2010; Maron and Moore,, 1993; Mnih et al.,, 2008; Bubeck et al.,, 2009). The prior work on MAB that is most relevant to ours is PAC learning for BAI and analysis of its sample complexity. We reduce the problem of BME to BAI and prove the lower bound on the sample complexity of BME using a technique known for BAI (Even-Dar et al.,, 2002). However, while this technique does not give tight lower bound for BAI (Mannor and Tsitsiklis,, 2004), we show that it gives tight lower bound for BME. Notice that the problem of estimating the best mean frequently appears in reinforcement learning (van Hasselt,, 2010) and machine learning (Kajino et al.,, 2023), but there the focus is on how best to estimate the best mean with a given set of samples (van Hasselt,, 2013), while our focus is on how best to collect samples to estimate the best mean.

3 Settings

The goal of mechanism design is to specify the rules of a game in a way that an outcome desired by the mechanism designer is achieved when rational players (i.e., players whose goal it is to maximize their individual utility) participate in that game (Jackson,, 2014; Shoham and Leyton-Brown,, 2009). Formally, let 𝒩{1,2,,N}\mathcal{N}\coloneqq\{1,2,\ldots,N\} be the set of players and 𝒪\mathcal{O} be the set of possible outcomes. For each player n𝒩n\in\mathcal{N}, let 𝒜n\mathcal{A}_{n} be the set of available actions and 𝒯n\mathcal{T}_{n} be the set of possible types. Let 𝒜𝒜1××𝒜N\mathcal{A}\coloneqq\mathcal{A}_{1}\times\ldots\times\mathcal{A}_{N} and 𝒯𝒯1××𝒯N\mathcal{T}\coloneqq\mathcal{T}_{1}\times\ldots\times\mathcal{T}_{N} be the corresponding product spaces. A mechanism μ:𝒜𝒪\mu:\mathcal{A}\to\mathcal{O} determines an outcome depending of the actions taken by the players. Let un:𝒪×𝒯nu_{n}:\mathcal{O}\times\mathcal{T}_{n}\to\mathbb{R} be the utility function of each player n𝒩n\in\mathcal{N}.

We consider Bayesian games where the players’ types follow a probability distribution that is known to all players and the mediator. Before selecting actions, the players know their own types but not the types of the other players. A strategy of each player n𝒩n\in\mathcal{N} is thus a function from 𝒯n\mathcal{T}_{n} to 𝒜n\mathcal{A}_{n}.

We assume that an outcome is determined by a social decision and payment; hence, a mechanism μ\mu consists of a decision rule and a payment rule. Let 𝒟\mathcal{D} be the set of possible social decisions. Given the actions of the players, the decision rule ϕ:𝒜𝒟\phi:\mathcal{A}\to\mathcal{D} determines a social decision, and the payment rule τ:𝒜𝒩\tau:\mathcal{A}\to\mathbb{R}^{\mathcal{N}} determines the amount of (possibly negative) payment to the mediator from each player. Let v:𝒟×(𝒯1𝒯N)v:\mathcal{D}\times(\mathcal{T}_{1}\cup\ldots\cup\mathcal{T}_{N})\to\mathbb{R} specify the value of a given social decision to the player of a given type. Then the utility of player ii when players take actions a𝒜a\in\mathcal{A} is

un(μ(a);tn)=un((ϕ(a),τ(a));tn)=v(ϕ(a);tn)τn(a).\displaystyle u_{n}(\mu(a);t_{n})=u_{n}((\phi(a),\tau(a));t_{n})=v(\phi(a);t_{n})-\tau_{n}(a). (1)

Throughout, we assume that 𝒩\mathcal{N}, 𝒟\mathcal{D}, and 𝒯n,n𝒩\mathcal{T}_{n},\forall n\in\mathcal{N} are finite sets.

Without loss of generality by the revelation principle (Shoham and Leyton-Brown,, 2009), we consider only direct mechanisms, where the action available to each player is to declare which type the player belongs to from the set of possible types (i.e., 𝒜n=𝒯n,n𝒩\mathcal{A}_{n}=\mathcal{T}_{n},\forall n\in\mathcal{N}). We will thus use 𝒯n\mathcal{T}_{n} for 𝒜n\mathcal{A}_{n}.

Then we seek to achieve the following four properties with our mechanisms. The first property is Dominant Strategy Incentive Compatibility (DSIC), which ensures that the optimal strategy of each player is to truthfully reveal its type regardless of the strategies of the other players. Formally, {fleqn}

[DSIC] v(ϕ(tn,tn);tn)τn(tn,tn)v(ϕ(t);tn)τn(t),t𝒯,tn𝒯n,n𝒩,\displaystyle\mbox{[DSIC] }v(\phi(t_{n},t_{-n}^{\prime});t_{n})-\tau_{n}(t_{n},t_{-n}^{\prime})\geq v(\phi(t^{\prime});t_{n})-\tau_{n}(t^{\prime}),\forall t^{\prime}\in\mathcal{T},\forall t_{n}\in\mathcal{T}_{n},\forall n\in\mathcal{N}, (2)

where the left-hand side represents the utility of the player having type tnt_{n} when it declares the same tnt_{n}, and the other players declare arbitrary types tnt_{-n}^{\prime}.

The second property is Decision Efficiency (DE), which requires that the mediator chooses the social decision that maximizes the total value to the players. With DSIC, we can assume that the players declare true types, and hence we can write DE as a condition on the decision rule: {fleqn}[10pt]

[DE]ϕ(t)argmaxd𝒟n𝒩v(d;tn)t𝒯.\displaystyle\mbox{[DE]}\qquad\phi(t)\in\mathop{\rm argmax}\limits_{d\in\mathcal{D}}\sum_{n\in\mathcal{N}}v(d;t_{n})\qquad\forall t\in\mathcal{T}. (3)

As the third property, we generalize individual rationality and require that the expected utility of each player is at least as large as a target value that can depend on its type. We refer to this property as θ\theta-IR. Again, assuming that players declare true types due to DSIC, we can write θ\theta-IR as follows: {fleqn}[10pt]

[θ-IR]𝔼[v(ϕ(t);tn)τn(t)tn]θ(tn)tn𝒯n,n𝒩,\displaystyle\mbox{[$\theta$-IR]}\qquad\mathbb{E}[v(\phi(t);t_{n})-\tau_{n}(t)\mid t_{n}]\geq\theta(t_{n})\qquad\forall t_{n}\in\mathcal{T}_{n},\forall n\in\mathcal{N}, (4)

where θ:𝒯1𝒯N\theta:\mathcal{T}_{1}\cup\ldots\cup\mathcal{T}_{N}\to\mathbb{R} determines the target expected utility for each type. Throughout (except in Section 6, where we discuss general MAB models), 𝔼\mathbb{E} denotes the expectation with respect to the probability distribution \mathbb{P} of types, which is the only probability that appears in our mechanisms.

The last property is a generalization of Budget Balance (BB), which we refer to as ρ\rho-WBB and ρ\rho-SBB. Specifically, ρ\rho-WBB requires that the expected revenue of the mediator is no less than a given constant ρ\rho\in\mathbb{R}, and ρ\rho-SBB requires that it is equal to ρ\rho. Again, assuming that the players declare true types due to DSIC, these properties can be written as follows: {fleqn}[10pt]

[ρ-WBB]n𝒩𝔼[τn(t)]ρ.[ρ-SBB]n𝒩𝔼[τn(t)]=ρ.\displaystyle\mbox{[$\rho$-WBB]}\qquad\sum_{n\in\mathcal{N}}\mathbb{E}\left[\tau_{n}(t)\right]\geq\rho.\qquad\qquad\qquad\mbox{[$\rho$-SBB]}\qquad\sum_{n\in\mathcal{N}}\mathbb{E}\left[\tau_{n}(t)\right]=\rho. (5)

While ρ\rho-SBB is stronger than ρ\rho-WBB, we will see that ρ\rho-SBB is satisfiable iff ρ\rho-WBB is satisfiable.

4 Optimization problem for automated mechanism design

Following Osogami et al., (2023), we seek to find optimal mechanisms in the class of VCG mechanisms. A VCG mechanism is specified by a pair (ϕ,h)(\phi^{\star},h). Specifically, after letting player take the actions of declaring their types t𝒯t\in\mathcal{T}, the mechanism first finds a social decision ϕ(t)\phi^{\star}(t) using a decision rule ϕ=ϕ\phi=\phi^{\star} that satisfies DE (3). It then determines the amount of payment from each player n𝒩n\in\mathcal{N} to the mediator by

τn(t)\displaystyle\tau_{n}(t) =hn(tn)m𝒩nv(ϕ(t);tm),\displaystyle=h_{n}(t_{-n})-\sum_{m\in\mathcal{N}_{-n}}v(\phi^{\star}(t);t_{m}), (6)

where we define 𝒩n𝒩{n}\mathcal{N}_{-n}\coloneqq\mathcal{N}\setminus\{n\}, and hn:𝒯nh_{n}:\mathcal{T}_{-n}\to\mathbb{R} is an arbitrary function of the types of the players other than nn and referred to as a pivot rule. The decision rule ϕ\phi^{\star} guarantees DE (3) by construction, and the payment rule (6) then guarantees DSIC (2) (Nisan,, 2007).

Our problem is now reduced to find the pivot rule, h={hn}n𝒩h=\{h_{n}\}_{n\in\mathcal{N}}, that minimizes the expected revenue of the mediator, while satisfying θ\theta-IR and ρ\rho-WBB. This may lead to satisfying ρ\rho-SBB if the revenue is maximally reduced. To represent this reduced problem, let

w(t)\displaystyle w^{\star}(t) n𝒩v(ϕ(t);tn)\displaystyle\coloneqq\sum_{n\in\mathcal{N}}v(\phi^{\star}(t);t_{n}) (7)

be the total value of the efficient social decision when the players have types tt. Then we can rewrite θ\theta-IR (for the player having type tnt_{n}) and ρ\rho-WBB as follows (see Appendix A.1 for full derivation):

𝔼[v(ϕ(t);tn)τn(t)tn]θ(tn)\displaystyle\mathbb{E}[v(\phi^{\star}(t);t_{n})-\tau_{n}(t)\mid t_{n}]\geq\theta(t_{n}) 𝔼[w(t)|tn]𝔼[hn(tn)|tn]θ(tn)\displaystyle\Longleftrightarrow\mathbb{E}\left[w^{\star}(t)\,\middle|\,t_{n}\right]-\mathbb{E}\left[h_{n}(t_{-n})\,\middle|\,t_{n}\right]\geq\theta(t_{n}) (8)
n𝒩𝔼[τn(t)]ρ\displaystyle\sum_{n\in\mathcal{N}}\mathbb{E}\left[\tau_{n}(t)\right]\geq\rho n𝒩𝔼[hn(tn)](N1)𝔼[w(t)]ρ.\displaystyle\Longleftrightarrow\sum_{n\in\mathcal{N}}\mathbb{E}\left[h_{n}(t_{-n})\right]-(N-1)\,\mathbb{E}\left[w^{\star}(t)\right]\geq\rho. (9)

Therefore, we arrive at the following linear program (LP):

minh\displaystyle\min_{h} n𝒩𝔼[hn(tn)]\displaystyle\qquad\sum_{n\in\mathcal{N}}\mathbb{E}[h_{n}(t_{-n})] (10)
s.t.\displaystyle\mathrm{s.t.} 𝔼[w(t)|tn]𝔼[hn(tn)|tn]θ(tn)tn𝒯n,n𝒩\displaystyle\qquad\mathbb{E}\left[w^{\star}(t)\,\middle|\,t_{n}\right]-\mathbb{E}\left[h_{n}(t_{-n})\,\middle|\,t_{n}\right]\geq\theta(t_{n})\qquad\forall t_{n}\in\mathcal{T}_{n},\forall n\in\mathcal{N} (11)
n𝒩𝔼[hn(tn)](N1)𝔼[w(t)]ρ.\displaystyle\qquad\sum_{n\in\mathcal{N}}\mathbb{E}\left[h_{n}(t_{-n})\right]-(N-1)\,\mathbb{E}\left[w^{\star}(t)\right]\geq\rho. (12)

The approach of Osogami et al., (2023) is to numerically solve this LP possibly with approximations. Since hn(tn)h_{n}(t_{-n}) is a variable for each tn𝒯nt_{-n}\in\mathcal{T}_{-n} and each n𝒩n\in\mathcal{N}, the LP has NKN1N\,K^{N-1} variables and NK+1N\,K+1 constraints, when each player has KK possible types. When this LP is feasible, let hh^{\star} be its optimal solution; then the VCG mechanism (ϕ,h)(\phi^{\star},h^{\star}) guarantees DSIC, DE, θ\theta-IR, and ρ\rho-WBB (formalized as Proposition 4 in Appendix A.1). Otherwise, no VCG mechanisms can guarantee them all. In the next section, we characterize exactly when the LP is feasible and provide analytical solutions to the LP.

5 Analytical solution to the optimization problem

We first establish a sufficient condition and a necessary condition for the LP to have feasible solutions. Note that complete proofs for all theoretical statements are provided in Appendix B.

Lemma 1.

The LP (10)-(12) is feasible if

n𝒩mintn𝒯n{𝔼[w(t)tn]θ(tn)}\displaystyle\sum_{n\in\mathcal{N}}\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\} (N1)𝔼[w(t)]+ρ.\displaystyle\geq(N-1)\,\mathbb{E}[w^{\star}(t)]+\rho. (13)
Lemma 2.

When types are independent (tmt_{m} and tnt_{n} are independent for any mnm\neq n under \mathbb{P}), the LP (10)-(12) is feasible only if (13) holds.

These two lemmas establish the following necessary and sufficient condition.

Corollary 1.

When types are independent, the LP (10)-(12) is feasible if and only if (13) holds.

While the LP is not necessarily feasible, one may choose θ\theta and ρ\rho in a way that it is guaranteed to be feasible. For example, the feasibility is guaranteed (i.e., (13) is satisfied) by setting

θ0andρ=[n𝒩mintn𝒯n𝔼[w(t)tn](N1)𝔼[w(t)]],\displaystyle\theta\equiv 0\qquad\mbox{and}\qquad\rho=\left[\sum_{n\in\mathcal{N}}\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t_{n}]-(N-1)\,\mathbb{E}[w^{\star}(t)]\right]^{-}, (14)

where [x]min{x,0}[x]^{-}\coloneqq\min\{x,0\} for xx\in\mathbb{R}. When ρ<0\rho<0, the mediator might get negative expected revenue, but the expected loss of the mediator is at most |ρ||\rho|. Appendix A.2 provides alternative ρ\rho and θ\theta that guarantee feasibility, but some of the players might incur negative expected utility.

Finally, we derive a class of optimal solutions to the LP when it is feasible.

Lemma 3.

A pivot rules is said to be constant if and only if, for each n𝒩n\in\mathcal{N}, there exists a constant ηn\eta_{n} such that hn(tn)=ηn,tn𝒯nh_{n}(t_{-n})=\eta_{n},\forall t_{-n}\in\mathcal{T}_{-n}. Let \mathcal{H} be the set of constant pivot rules that satisfy:

ηn\displaystyle\eta_{n} =mintn𝒯n{𝔼[w(t)tn]θ(tn)}δnn𝒩\displaystyle=\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\}-\delta_{n}\qquad\forall n\in\mathcal{N} (15)

where δ\delta lies on the following simplex:

δn\displaystyle\delta_{n} 0n𝒩\displaystyle\geq 0\qquad\forall n\in\mathcal{N} (16)
n𝒩δn\displaystyle\sum_{n\in\mathcal{N}}\delta_{n} =n𝒩mintn𝒯n{𝔼[w(t)tn]θ(tn)}(N1)𝔼[w(t)]ρ.\displaystyle=\sum_{n\in\mathcal{N}}\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\}-(N-1)\,\mathbb{E}[w^{\star}(t)]-\rho. (17)

Then, when (13) holds, \mathcal{H} is nonempty, and any hh\in\mathcal{H} is an optimal solution to the LP (10)-(12).

When types are independent, the condition (13) is necessary for the existence of a feasible solution; hence, we do not lose optimality by considering only the solutions in \mathcal{H}. On the other hand, when types are dependent, the condition (13) may still be satisfied, where the solutions in \mathcal{H} remain optimal. However, dependent types no longer necessitate (13) to satisfy (11)-(12); hence the space of constant pivot rules may not suffice to find an optimal solution if (13) does not hold.

Proposition 1.

The LP (10)-(12) may be feasible even if (13) is violated (when types are dependent).

Corollary 2.

Feasible solutions to the LP (10)-(12) include at least one constant pivot rule (defined in Lemma 3) if and only if (13) holds (whether types are dependent or not).

In deriving the optimal solutions, we have substantially reduced the essential number of variables (from NTN1N\,T^{N-1} to NN when each player has TT types). Our approach can therefore not only find but also represent or store solutions exponentially more efficiently than Osogami et al., (2023).

It turns out that the solutions in \mathcal{H} not only satisfy ρ\rho-WBB but also ρ\rho-SBB (in addition to DE, DSIC, and θ\theta-IR) regardless of whether the types are independent or not:

Corollary 3.

Any VCG mechanism given with a pivot rule in \mathcal{H} satisfies ρ\rho-SBB.

When the LP is infeasible, we can construct a mechanism that satisfies one of ρ\rho-SBB and θ\theta-IR (in addition to DE and DSIC) regardless of whether the types are independent or not:

Corollary 4.

Any VCG mechanism with a pivot rule given by (15) and (17) satisfies ρ\rho-SBB.

Corollary 5.

Any VCG mechanism with a pivot rule given by (15) and (16) satisfies θ\theta-IR for any tn𝒯nt_{n}\in\mathcal{T}_{n} and n𝒩n\in\mathcal{N}.

For example, for any tn𝒯nt_{n}\in\mathcal{T}_{n} and n𝒩n\in\mathcal{N}, the following pivot rule always satisfies θ\theta-IR:

ηn\displaystyle\eta_{n} =mintn𝒯n{𝔼[w(t)tn]θ(tn)}max{δ,0},\displaystyle=\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\}-\max\{\delta,0\}, (18)

where we define (see also (58) in the appendix)

δ1N(n𝒩mintn𝒯n{𝔼[w(t)tn]θ(tn)}(N1)𝔼[w(t)]ρ).\displaystyle\delta\coloneqq\frac{1}{N}\bigg{(}\sum_{n\in\mathcal{N}}\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\}-(N-1)\,\mathbb{E}[w^{\star}(t)]-\rho\bigg{)}. (19)

So far, we have analytically derived the class of optimal solutions for the LP. Although these analytical solutions substantially reduce the computational cost of solving the LP compared to the numerical solutions in Osogami et al., (2023), they still need to evaluate

mintn𝒯n{𝔼[w(t)tn]θ(tn)}\displaystyle\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\} (20)

for each n𝒩n\in\mathcal{N}. Since 𝔼\mathbb{E} is the expectation with respect to the distribution \mathbb{P} over 𝒯\mathcal{T}, this would require evaluating w(t)w^{\star}(t) for all t𝒯t\in\mathcal{T}. Recall that w(t)w^{\star}(t) defined with (7) is the total value of the efficient decision ϕ(t)\phi^{\star}(t), which is given by a solution to an optimization problem (3). Without any structure in 𝒟\mathcal{D} and vv, this would require evaluating the total value for all decisions in 𝒟\mathcal{D}.

6 Online learning for evaluating the analytical solution

To alleviate the computational complexity associated with (20), we propose a learning approach. A key observation is that the problem of estimating (20) can be considered as a variant of a multi-armed bandit (MAB) problem whose objective is to estimate the mean reward of the best arm. Specifically, the MAB gives the reward θ(tn)w(t)\theta(t_{n})-w^{\star}(t) when we pull an arm tn𝒯nt_{n}\in\mathcal{T}_{n}, where tt is a sample from the conditional distribution [tn]\mathbb{P}[\cdot\mid t_{n}]. Following the relevant prior work on MAB (Even-Dar et al.,, 2002, 2006; Hassidim et al.,, 2020; Mannor and Tsitsiklis,, 2004), we maximize reward in this section.

Since we assume that 𝒩\mathcal{N}, 𝒟\mathcal{D}, and 𝒯n,n𝒩\mathcal{T}_{n},\forall n\in\mathcal{N} are finite, there exist constants, θ¯\bar{\theta} and v¯\bar{v}, such that |θ(t)|θ¯|\theta(t^{\prime})|\leq\bar{\theta} and |v(d;t)|v¯,d𝒟,tn𝒩𝒯n|v(d;t^{\prime})|\leq\bar{v},\forall d\in\mathcal{D},\forall t^{\prime}\in\cup_{n\in\mathcal{N}}\mathcal{T}_{n}. Then we can also bound |θ(tn)w(t)|θ¯+Nv¯,t𝒯,n𝒩|\theta(t_{n})-w^{\star}(t)|\leq\bar{\theta}+N\,\bar{v},\forall t\in\mathcal{T},\forall n\in\mathcal{N}. Namely, the reward in the MAB is bounded. We assume that we know the bounds on the reward and can scale it such that the scaled reward is in [0,1][0,1] surely.

We also assume that we have access to an arbitrary size of the sample that is independent and identically distributed (i.i.d.) according to [tn]\mathbb{P}[\cdot\mid t_{n}] for any tn𝒯n,n𝒩t_{n}\in\mathcal{T}_{n},n\in\mathcal{N}. When players have independent types, such sample can be easily constructed as long as we have access to i.i.d. sample {t(i)}i=1,2,\{t^{(i)}\}_{i=1,2,\ldots} from 𝒯\mathcal{T}, because {(tn,tn(i))}i\{(t_{n},t_{-n}^{(i)})\}_{i} is the sample from [tn]\mathbb{P}[\cdot\mid t_{n}] for any tn𝒯n,n𝒩t_{n}\in\mathcal{T}_{n},n\in\mathcal{N}.

Consider the general KK-armed bandit where the reward of each arm is bounded in [0,1][0,1]. For each k{1,2,,K}k\in\{1,2,\ldots,K\}, let μk\mu_{k} be the true mean of arm kk. Let μmaxkμk\mu_{\star}\coloneqq\max_{k}\mu_{k} be the best mean-reward, which we want to estimate. We say that the sample complexity of an algorithm for a MAB is TT if the sample size needed by the algorithm is bounded by TT (i.e., arms are pulled at most TT times).

A standard Probably Approximately Correct (PAC) algorithm for MAB returns an ε\varepsilon-optimal arm with probability at least 1δ1-\delta for given ε,δ\varepsilon,\delta (Even-Dar et al.,, 2006; Hassidim et al.,, 2020). We will use the following definition:

Definition 1 ((ε,δ\varepsilon,\delta)-PAC Best Arm Identifier (Even-Dar et al.,, 2006)).

For ε,δ>0\varepsilon,\delta>0, we say that an algorithm is (ε,δ\varepsilon,\delta)-PAC Best Arm Identifier (BAI) if the arm I^\hat{I} identified by the algorithm satisfies

Pr(μI^με)1δ.\displaystyle\Pr\left(\mu_{\hat{I}}\geq\mu_{\star}-\varepsilon\right)\geq 1-\delta. (21)

Definition 1 is different from what we need to evaluate (20). Formally, we need

Definition 2 ((ε,δ\varepsilon,\delta)-PAC Best Mean Estimator).

For ε,δ>0\varepsilon,\delta>0, we say that an algorithm is (ε,δ\varepsilon,\delta)-PAC Best Mean Estimator (BME) if the best mean μ^\hat{\mu} estimated by the algorithm satisfies

Pr(|μ^μ|ε)1δ.\displaystyle\Pr\left(|\hat{\mu}-\mu_{\star}|\leq\varepsilon\right)\geq 1-\delta. (22)

BAI and BME are related but different. For example, consider the case where the best arm has large variance and μ=1/2\mu_{\star}=1/2, and the other arms always give zero reward μn=0,n\mu_{n}=0,\forall n\neq\star. Then relatively small sample would suffice for BAI due to the large gap μμn=1/2,n\mu_{\star}-\mu_{n}=1/2,\forall n\neq\star, while BME would require relatively large sample due to the large variance of the best arm \star. As another example, consider the case where we have many arms whose rewards follow Bernoulli distributions. Suppose that half of the arms have an expected value of 1, and the other half have an expected value of 1(3/2)ε1-(3/2)\,\varepsilon. Then by pulling sufficiently many arms (once for each arm), we can estimate that the best mean is at least 1ε1-\varepsilon with high probability (by Hoeffding’s inequality), but BAI would require pulling arms sufficiently many times until we pull one of the best arms sufficiently many times (to be able to say that this particular arm has an expected value at least 1ε1-\varepsilon).

For (ε,δ\varepsilon,\delta)-PAC BAI, the following lower and upper bounds on the sample complexity are known:

Proposition 2 (Mannor and Tsitsiklis, (2004)).

There exists a (ε,δ\varepsilon,\delta)-PAC BAI with sample complexity O((K/ε2)log(1/δ))O((K/\varepsilon^{2})\,\log(1/\delta)), and any (ε,δ\varepsilon,\delta)-PAC BAI must have the sample complexity at least Ω((K/ε2)log(1/δ))\Omega((K/\varepsilon^{2})\,\log(1/\delta)).

We establish the same lower and upper bounds for (ε,δ\varepsilon,\delta)-PAC BME:

Theorem 1.

There exists a (ε,δ\varepsilon,\delta)-PAC BME with sample complexity O((K/ε2)log(1/δ))O((K/\varepsilon^{2})\,\log(1/\delta)), and any (ε,δ\varepsilon,\delta)-PAC BME must have the sample complexity at least Ω((K/ε2)log(1/δ))\Omega((K/\varepsilon^{2})\,\log(1/\delta)).

Our upper bound is established by reducing BME to BAI. Suppose that we have access to an arbitrary (ε,δ)(\varepsilon,\delta)-PAC BAI with sample complexity MM (Algorithm 2 in Appendix A.3). We can construct a (32ε,2δ)(\frac{3}{2}\varepsilon,2\delta)-PAC BME by first running the (ε,δ)(\varepsilon,\delta)-PAC BAI and then sampling from the arm I^\hat{I} that is identified as the best until we have a sufficient number, mm^{\star}, of samples from I^\hat{I} (Algorithm 3 in Appendix A.3). When mm^{\star} is appropriately selected, the following lemma holds:

Lemma 4.

When a (ε,δ)(\varepsilon,\delta)-PAC BAI with sample complexity MM is used, Algorithm 3 has sample complexity M+mM+m^{\star}, where m(2/ε2)log(1.22/δ)m^{\star}\coloneqq\left\lceil(2/\varepsilon^{2})\log(1.22/\delta)\right\rceil, and returns μ^I^\hat{\mu}_{\hat{I}} that satisfies

Pr(|μ^I^μ|>32ε)2δ.\displaystyle\Pr\left(|\hat{\mu}_{\hat{I}}-\mu_{\star}|>\frac{3}{2}\varepsilon\right)\leq 2\,\delta. (23)

By Proposition 2, this establishes the upper bound on the sample complexity in Theorem 1.

Our lower bound is established by showing that an arbitrary (ε,δ)(\varepsilon,\delta)-PAC BME can be used to solve the problem of identifying whether a given coin is negatively or positively biased (precisely, ε\varepsilon-Biased Coin Problem of Definition 3 in Appendix A.3), for which any algorithm is known to require expected sample complexity at least Ω((1/ε2)log(1/δ))\Omega((1/\varepsilon^{2})\,\log(1/\delta)) to solve correctly with probability at least 1δ1-\delta (Chernoff,, 1972; Even-Dar et al.,, 2002) (see Lemma 7 in Appendix A.3). The following lemma, together with Lemma 7, establishes the lower bound on the sample complexity in Theorem 1:

Lemma 5.

If there exists an (ε/2,δ/2)(\varepsilon/2,\delta/2)-PAC BME with sample complexity MM for KK-armed bandit, then there also exists an algorithm, having expected sample complexity M/KM/K, that can solve the ε\varepsilon-Biased Coin Problem correctly with probability at least 1δ1-\delta.

We prove Lemma 5 by reducing the ε\varepsilon-Biased Coin Problem to BME. This proof technique was also used in Even-Dar et al., (2002) to prove a lower bound on the sample complexity of BAI. What is interesting, however, is that the lower bound in Even-Dar et al., (2002) is not tight when δ<1/K\delta<1/K, and a tight lower bound is established by a different technique in Mannor and Tsitsiklis, (2004). On the other hand, our proof gives a tight lower bound on the sample complexity of BME. In Appendix A.3, we further discuss where this difference in the derived lower bounds comes from.

Finally, we connect the results on BME back to our mechanism in Section 5 and provide a guarantee on the properties of the mechanism when the term (20) is estimated with BME. To this end, we define the terms involving expectations in Lemma 3 as follows:

κn(θ)\displaystyle\kappa_{n}(\theta) mintn𝒯N{𝔼[wtn]θ(tn)}\displaystyle\coloneqq\min_{t_{n}\in\mathcal{T}_{N}}\left\{\mathop{\mathbb{E}}\limits[w^{\star}\mid t_{n}]-\theta(t_{n})\right\} (24)
λ(ρ)\displaystyle\lambda(\rho) 𝔼[w(t)]+ρ/(N1).\displaystyle\coloneqq\mathop{\mathbb{E}}\limits[w^{\star}(t)]+\rho/(N-1). (25)

Then, recalling that NN is the number of players, we have the following lemma:

Lemma 6.

Let κ~n(θ)\tilde{\kappa}_{n}(\theta) for n𝒩n\in\mathcal{N} and λ~(ρ)\tilde{\lambda}(\rho) be independent estimates of κn(θ)\kappa_{n}(\theta) and λ(ρ)\lambda(\rho) respectively given by an (ε,δ)(\varepsilon^{\prime},\delta^{\prime})-PAC Best Mean Estimator and a standard (ε′′,δ)(\varepsilon^{\prime\prime},\delta^{\prime})-PAC estimator of expectation. Also, let d~d(κ~(θ),λ~(ρ),ε′′′,ε′′′′)\tilde{d}\coloneqq d(\tilde{\kappa}(\theta),\tilde{\lambda}(\rho),\varepsilon^{\prime\prime\prime},\varepsilon^{\prime\prime\prime\prime}) be a point on the following simplex (here, we change the notation from δ\delta in Lemma 3 to d~\tilde{d} to avoid confusion):

d~n\displaystyle\tilde{d}_{n} ε′′′,n𝒩\displaystyle\geq\varepsilon^{\prime\prime\prime},\forall n\in\mathcal{N} (26)
n𝒩d~n\displaystyle\sum_{n\in\mathcal{N}}\tilde{d}_{n} =n𝒩κ~n(θ)(N1)(λ~(θ)+ε′′′′)\displaystyle=\sum_{n\in\mathcal{N}}\tilde{\kappa}_{n}(\theta)-(N-1)\,(\tilde{\lambda}(\theta)+\varepsilon^{\prime\prime\prime\prime}) (27)

Then the VCG mechanism with the constant pivot rule hn(tn)=ηn=κ~n(θ)d~nh_{n}(t_{-n})=\eta_{n}=\tilde{\kappa}_{n}(\theta)-\tilde{d}_{n} satisfies (θ(εε′′′))(\theta-(\varepsilon^{\prime}-\varepsilon^{\prime\prime\prime}))-IR and (ρ(N1)(ε′′ε′′′′))(\rho-(N-1)\,(\varepsilon^{\prime\prime}-\varepsilon^{\prime\prime\prime\prime}))-WBB with probability (1δ)N+1(1-\delta^{\prime})^{N+1}.

Notice that the sufficient condition of Lemma 1 states that the simplex in Lemma 3 is nonempty. Analogously, when the simplex in Lemma 6 is empty, we cannot provide the solution that guarantees the properties stated in Lemma 6. Since DSIC and DE remain satisfied regardless of whether κn\kappa_{n} for n𝒩n\in\mathcal{N} and λ\lambda are estimated or exactly computed, Lemma 6 immediately establishes the following theorem by appropriately choosing the parameters:

Theorem 2.

In Lemma 6, let ε′′′=ε\varepsilon^{\prime\prime\prime}=\varepsilon^{\prime}, ε′′′′=ε′′\varepsilon^{\prime\prime\prime\prime}=\varepsilon^{\prime\prime}, and δ=1(1δ)1/(N+1)\delta^{\prime}=1-(1-\delta)^{1/(N+1)}. Then the VCG mechanism with the constant pivot rule hn(tn)=κ~n(θ)d~nh_{n}(t_{-n})=\tilde{\kappa}_{n}(\theta)-\tilde{d}_{n} satisfies DSIC, DE, θ\theta-IR, and ρ\rho-WBB with probability 1δ1-\delta.

Recall that the exact computation of our mechanism in Lemma 3 requires evaluating w(t)w^{\star}(t) for all t𝒯t\in\mathcal{T}, whose size grows exponentially with the number of players, NN. Our BME algorithm reduces this requirement to O(NlogN)O(N\,\log N), as is formally proved in the following proposition:

Proposition 3.

The sample complexity to learn the constant pivot rule in Theorem 2 is O((NK/ε2)log(N/δ))O((N\,K/\varepsilon^{2})\,\log(N/\delta)), where N=|𝒩|N=|\mathcal{N}| is the number of players, and K=maxn𝒩𝒯nK=\max_{n\in\mathcal{N}}\mathcal{T}_{n} is the maximum number of types of each player.

7 Numerical experiments

We conduct several numerical experiments to validate the effectiveness of the proposed approach and to understand its limitations. Specifically, we design our experiments to investigate the following questions. i) BME studied in Section 6 has asymptotically optimal sample complexity, but how well can we estimate the best mean when there are only a moderate number of arms? ii) How much can BME reduce the number of times w(t)w^{\star}(t) given by (7) is evaluated? iii) How well θ\theta-IR and ρ\rho-SBB are satisfied when (20) is estimated by BME rather than calculated exactly? In this section, we only provide a brief overview of our experiments; for full details, see Appendix A.4. In particular, we find the key empirical property of BME that it can reduce the computational complexity by many orders of magnitude when the number of players is moderate (10\sim 10) to large but is relatively less effective against the increased number of types.

For question i), we compare Successive Elimination for BAI (SE-BAI), which is known to perform well for a moderate number of arms (Hassidim et al.,, 2020; Even-Dar et al.,, 2006), against the corresponding algorithm for BME (SE-BME) and summarize the results in Appendix A.4.1. Overall, we find that SE-BME generally requires smaller sample size than SE-BAI, except when there are only a few arms and their means have large gaps. The efficiency of SE-BAI for this case makes intuitive sense, since the best arm can be identified without estimating the means with high accuracy.

Refer to caption

(a) sample path

Refer to caption

(b) sample size against NN

Refer to caption

(c) sample size against TT

Figure 1: (a) Representative sample path that shows the estimated values of 𝔼[w(t)tn]\mathbb{E}[w^{\star}(t)\mid t_{n}] for tn𝒯nt_{n}\in\mathcal{T}_{n} against sample size used by a (0.25,0.1)(0.25,0.1)-PAC BME with Successive Elimination (SE-BME; Algorithm 4) for 8 players, each with 8 types. (b)-(c) The unique sample size (the number of unique tt which w(t)w^{\star}(t) is evaluated with) required by the exact computation of mintn𝒯n𝔼[w(t)tn],n𝒩\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t_{n}],\forall n\in\mathcal{N} (dashed curves) and by SE-BME (solid curves).

For question ii), we quantitatively validate the effectiveness of SE-BME in reducing the number of times w(t)w^{\star}(t) is computed when we evaluate (20) in a setting of mechanism design, as is detailed in Appendix A.4.2. Figure 1 summarizes the results. Panel (a) shows that the means close to the minimum value are estimated with high accuracy, while others are eliminated by SE-BME after a small number of samples. Panels (b)-(c) show that SE-BME (solid curves) evaluates w(t)w^{\star}(t) by orders of magnitude smaller number of times than what is required by exact computation (dashed curves). The effectiveness of SE-BME is particularly striking when NN (the number of players) grows. While exact computation evaluates w(t)w^{\star}(t) exponentially many times as NN grows, SE-BME has only linear dependency on NN. This insensitivity of the sample complexity of SE-BME to NN makes intuitive sense, since NN only affects the distribution of the reward and keeps the number of arms unchanged.

IR is guaranteed with (57) and (18)

SBB is guaranteed with (57) and (58)

Refer to caption

(a) players’ utility

Refer to caption

(b) mediator’s revenue

Refer to caption

(c) players’ utility

Refer to caption

(d) mediator’s revenue

Figure 2: The red dot shows expected utility in (a) and (c) and expected revenue in (b) and (d), when analytical solutions are evaluated exactly (horizontal) or estimated with (0.25,0.1)(0.25,0.1)-PAC SE-BME (vertical) for environments with 88 players, each having 88 possible types. The analytical solution guarantees IR in (a) and (b) and SBB in (c) and (d). Results are plotted for 10 random seeds.

For question iii), we quantitatively evaluate how well θ\theta-IR and ρ\rho-SBB are guaranteed when we estimate (20) with SE-BME, where we continue to use the same setting of mechanism design. In Figure 2, we study two analytical solutions: one guarantees to satisfy θ\theta-IR (with θ0\theta\equiv 0), and the other guarantees to satisfy ρ\rho-SBB (with ρ=0\rho=0). The red dots show the expected utility of the players in (a) and (c) as well as the expected revenue of the mediator in (b) and (d), when (20) is evaluated either exactly (horizontal axes) or estimated with SE-BME (vertical axes). Overall, the expected revenue of the mediator is more sensitive than the expected utility of the players to the error in the estimation of (20), because (12) involves the summation n𝒩ηn\sum_{n\in\mathcal{N}}\eta_{n}, while (11) only involves ηn\eta_{n} for a single n𝒩n\in\mathcal{N}. We can however ensure that ρ\rho-WBB (instead of SBB) is satisfied with high probability by replacing the ρ\rho with a ρ>ρ\rho^{\prime}>\rho when we compute ηn\eta_{n} after (20) is estimated. As we show and explain with Figure 8 in Appendix A.4.3, this will simply shift the dots in the figure, and we can guarantee ρ\rho-WBB with high probability with an appropriate selection of ρ\rho^{\prime}.

We have run all of the experiments on a single core with at most 66 GB memory without GPUs in a cloud environment, as is detailed in Appendix A.4.4. The associated source code is submitted as a supplementary material and will be open-sourced upon acceptance.

8 Conclusion

We have analytically derived optimal solutions for the LP that gives mechanisms that guarantee the desired properties of DE, DSIC, ρ\rho-SBB, and θ\theta-IR. When there are NN players, each with TT possible types, the LP involves NTN1N\,T^{N-1} variables, while our analytical solutions are represented by only NN essential variables. While Osogami et al., (2023) numerically solves this LP only for N=T=2N=T=2, we have exactly evaluated our analytical solutions for N=T=8N=T=8 (see Figure 1). The analytical solution, however, involves a term whose exact evaluation requires finding efficient social decisions TNT^{N} times. We have modeled the problem of evaluating this term as best mean estimation in multi-armed bandit, proposed a PAC estimator, and proved its asymptotic optimality by establishing a lower bound on the sample complexity. Our experiments show that our PAC estimator enables finding mechanism for N=128N=128 and T=8T=8 with a guarantee on the desired properties.

The proposed approach makes a major advancement in the field and can positively impact society by providing guarantees on desired properties for large environments, which existing approaches are unable to handle. However, it presents certain limitations and potential challenges, which inspire several directions for further research. Regarding limitations, one should keep in mind that our mechanisms guarantee ρ\rho-SBB and θ\theta-IR only in expectation, and that the our approaches have only been applied to the environments up to 128 players and up to 16 types in our experiments. Regarding societal impacts, our approach might result in the mechanisms that are unfair to some of the players, because DE does not mean that all of the players are treated fairly. These limitations and societal impacts are further discussed in Appendix A.5.

References

  • Audibert et al., (2010) Audibert, J., Bubeck, S., and Munos, R. (2010). Best arm identification in multi-armed bandits. In Kalai, A. T. and Mohri, M., editors, Proceedings of the 23rd Conference on Learning Theory, pages 41–53. Omnipress.
  • Auer et al., (1995) Auer, P., Cesa-Bianchi, N., Freund, Y., and Schapire, R. (1995). Gambling in a rigged casino: The adversarial multi-armed bandit problem. In Proceedings of IEEE 36th Annual Foundations of Computer Science, pages 322–331.
  • Auer et al., (2002) Auer, P., Cesa-Bianchi, N., and P., F. (2002). Finite-time analysis of the multiarmed bandit problem. Machine Learning, 47:235––256.
  • Bailey, (1997) Bailey, M. J. (1997). The demand revealing process: To distribute the surplus. Public Choice, 91(2):107–126.
  • Balcan et al., (2016) Balcan, M.-F., Sandholm, T., and Vitercik, E. (2016). Sample complexity of automated mechanism design. In Proceedings of the 30th International Conference on Neural Information Processing Systems, NIPS’16, page 2091–2099, Red Hook, NY, USA. Curran Associates Inc.
  • Bubeck et al., (2009) Bubeck, S., Munos, R., and Stoltz, G. (2009). Pure exploration in multi-armed bandits problems. In Gavaldà, R., Lugosi, G., Zeugmann, T., and Zilles, S., editors, Algorithmic Learning Theory, pages 23–37, Berlin, Heidelberg. Springer Berlin Heidelberg.
  • Cavallo, (2006) Cavallo, R. (2006). Optimal decision-making with minimal waste: Strategyproof redistribution of VCG payments. In Proceedings of the Fifth International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS ’06, page 882–889, New York, NY, USA. Association for Computing Machinery.
  • Chernoff, (1972) Chernoff, H. (1972). Sequential Analysis and Optimal Design. CBMS-NSF Regional Conference Series in Applied Mathematics, Series Number 8. Society for Industrial and Applied Mathematics.
  • Chichin et al., (2017) Chichin, S., Vo, Q. B., and Kowalczyk, R. (2017). Towards efficient and truthful market mechanisms for double-sided cloud markets. IEEE Transactions on Services Computing, 10(1):37–51.
  • Conitzer and Sandholm, (2002) Conitzer, V. and Sandholm, T. (2002). Complexity of mechanism design. In Proceedings of the 18th Conference on Uncertainty in Artificial Intelligence, page 103–110.
  • Curry et al., (2020) Curry, M., Chiang, P.-Y., Goldstein, T., and Dickerson, J. (2020). Certifying strategyproof auction networks. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 4987–4998. Curran Associates, Inc.
  • Duetting et al., (2019) Duetting, P., Feng, Z., Narasimhan, H., Parkes, D., and Ravindranath, S. S. (2019). Optimal auctions through deep learning. In Chaudhuri, K. and Salakhutdinov, R., editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 1706–1715. PMLR.
  • Dufton et al., (2021) Dufton, L., Naroditskiy, V., Polukarov, M., and Jennings, N. (2021). Optimizing payments in dominant-strategy mechanisms for multi-parameter domains. Proceedings of the AAAI Conference on Artificial Intelligence, 26(1):1347–1354.
  • Even-Dar et al., (2002) Even-Dar, E., Mannor, S., and Mansour, Y. (2002). PAC bounds for Multi-armed bandit and Markov decision processes. In Kivinen, J. and Sloan, R. H., editors, Computational Learning Theory, pages 255–270, Berlin, Heidelberg. Springer Berlin Heidelberg.
  • Even-Dar et al., (2006) Even-Dar, E., Mannor, S., and Mansour, Y. (2006). Action elimination and stopping conditions for the multi-armed bandit and reinforcement learning problems. Journal of Machine Learning Research, 7(39):1079–1105.
  • Green and Laffont, (1977) Green, J. R. and Laffont, J. J. (1977). Characterization of satisfactory mechanism for the revelation of preferences for public goods. Econometrica, 45:427–438.
  • Green and Laffont, (1979) Green, J. R. and Laffont, J.-J. (1979). Incentives in Public Decision Making. North-Holland Publishing Company, Amsterdam.
  • Gujar and Narahari, (2011) Gujar, S. and Narahari, Y. (2011). Redistribution mechanisms for assignment of heterogeneous objects. Journal of Artificial Intelligence Research, 41(2):131–154.
  • Guo, (2011) Guo, M. (2011). VCG redistribution with gross substitutes. Proceedings of the AAAI Conference on Artificial Intelligence, 25(1):675–680.
  • Guo, (2012) Guo, M. (2012). Worst-case optimal redistribution of VCG payments in heterogeneous-item auctions with unit demand. In Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems - Volume 2, AAMAS ’12, page 745–752, Richland, SC. International Foundation for Autonomous Agents and Multiagent Systems.
  • Guo and Conitzer, (2007) Guo, M. and Conitzer, V. (2007). Worst-case optimal redistribution of VCG payments. In Proceedings of the 8th ACM Conference on Electronic Commerce, EC ’07, page 30–39, New York, NY, USA. Association for Computing Machinery.
  • Guo and Conitzer, (2009) Guo, M. and Conitzer, V. (2009). Worst-case optimal redistribution of VCG payments in multi-unit auctions. Games and Economic Behavior, 67(1):69–98. Special Section of Games and Economic Behavior Dedicated to the 8th ACM Conference on Electronic Commerce.
  • Hassidim et al., (2020) Hassidim, A., Kupfer, R., and Singer, Y. (2020). An optimal elimination algorithm for learning a best arm. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H., editors, Advances in Neural Information Processing Systems, volume 33, pages 10788–10798. Curran Associates, Inc.
  • Hatfield et al., (2013) Hatfield, J. W., Kominers, S. D., Nichifor, A., Ostrovsky, M., and Westkamp, A. (2013). Stability and competitive equilibrium in trading networks. Journal of Political Economy, 121(5):966–1005.
  • Hobbs et al., (2000) Hobbs, B. F., Rothkopf, M. H., Hyde, L. C., and O’Neill, R. P. (2000). Evaluation of a truthful revelation auction in the context of energy markets with nonconcave benefits. Journal of Regulatory Economics, 18:5–32.
  • Ivanov et al., (2022) Ivanov, D., Safiulin, I., Filippov, I., and Balabaeva, K. (2022). Optimal-er auctions through attention. In Oh, A. H., Agarwal, A., Belgrave, D., and Cho, K., editors, Advances in Neural Information Processing Systems.
  • Jackson, (2014) Jackson, M. O. (2014). Mechanism theory. Available at SSRN: http://dx.doi.org/10.2139/ssrn.2542983.
  • Kajino et al., (2023) Kajino, H., Miyaguchi, K., and Osogami, T. (2023). Biases in evaluation of molecular optimization methods and bias reduction strategies. In Krause, A., Brunskill, E., Cho, K., Engelhardt, B., Sabato, S., and Scarlett, J., editors, Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pages 15567–15585. PMLR.
  • Kumar et al., (2018) Kumar, D., Baranwal, G., Raza, Z., and Vidyarthi, D. P. (2018). A truthful combinatorial double auction-based marketplace mechanism for cloud computing. Journal of Systems and Software, 140:91–108.
  • Lattimore and Szepesvári, (2020) Lattimore, T. and Szepesvári, C. (2020). Bandit Algorithms. Cambridge University Press.
  • Manisha et al., (2018) Manisha, P., Jawahar, C. V., and Gujar, S. (2018). Learning optimal redistribution mechanisms through neural networks. In Proceedings of the 17th International Conference on Autonomous Agents and MultiAgent Systems, page 345–353.
  • Mannor and Tsitsiklis, (2004) Mannor, S. and Tsitsiklis, J. N. (2004). The sample complexity of exploration in the multi-armed bandit problem. Journal of Machine Learning Research, 5:623–648.
  • Maron and Moore, (1993) Maron, O. and Moore, A. (1993). Hoeffding races: Accelerating model selection search for classification and function approximation. In Cowan, J., Tesauro, G., and Alspector, J., editors, Advances in Neural Information Processing Systems, volume 6. Morgan-Kaufmann.
  • Mnih et al., (2008) Mnih, V., Szepesvári, C., and Audibert, J.-Y. (2008). Empirical Bernstein stopping. In Proceedings of the 25th International Conference on Machine Learning, ICML ’08, page 672–679, New York, NY, USA. Association for Computing Machinery.
  • Morgenstern and Roughgarden, (2015) Morgenstern, J. H. and Roughgarden, T. (2015). On the pseudo-dimension of nearly optimal auctions. In Cortes, C., Lawrence, N., Lee, D., Sugiyama, M., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 28. Curran Associates, Inc.
  • Moulin, (2009) Moulin, H. (2009). Almost budget-balanced VCG mechanisms to assign multiple objects. Journal of Economic Theory, 144(1):96–119.
  • Myerson, (1981) Myerson, R. B. (1981). Optimal auction design. Mathematics of Operations Research, 6:58–73.
  • Myerson and Satterthwaite, (1983) Myerson, R. B. and Satterthwaite, M. A. (1983). Efficient mechanisms for bilateral trading. Journal of Economic Theory, 29(2):265–281.
  • Nisan, (2007) Nisan, N. (2007). Introduction to mechanism design (for computer scientists). In Nisan, N., Roughgarden, T., Tardos, E., and Vazirani, V. V., editors, Algorithmic Game Theory, page 209–242. Cambridge University Press.
  • Osogami et al., (2023) Osogami, T., Wasserkrug, S., and Shamash, E. S. (2023). Learning efficient truthful mechanisms for trading networks. In Elkind, E., editor, Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence, IJCAI-23, pages 2862–2869. International Joint Conferences on Artificial Intelligence Organization. Main Track.
  • Rahme et al., (2021) Rahme, J., Jelassi, S., and Weinberg, S. M. (2021). Auction learning as a two-player game. In International Conference on Learning Representations.
  • Sandholm, (2003) Sandholm, T. (2003). Automated mechanism design: A new application area for search algorithms. In Rossi, F., editor, Principles and Practice of Constraint Programming – CP 2003, pages 19–36, Berlin, Heidelberg. Springer Berlin Heidelberg.
  • Shoham and Leyton-Brown, (2009) Shoham, Y. and Leyton-Brown, K. (2009). Multiagent Systems: Algorithmic, Game-Theoretic, and Logical Foundations. Cambridge University Press.
  • Stößer et al., (2010) Stößer, J., Neumann, D., and Weinhardt, C. (2010). Market-based pricing in grids: On strategic manipulation and computational cost. European Journal of Operational Research, 203(2):464–475.
  • Syrgkanis, (2017) Syrgkanis, V. (2017). A sample complexity measure with applications to learning optimal auctions. In Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., editors, Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  • Tacchetti et al., (2022) Tacchetti, A., Strouse, D., Garnelo, M., Graepel, T., and Bachrach, Y. (2022). Learning truthful, efficient, and welfare maximizing auction rules. In ICLR 2022 Workshop on Gamification and Multiagent Solutions.
  • van Hasselt, (2010) van Hasselt, H. (2010). Double Q-learning. In Lafferty, J., Williams, C., Shawe-Taylor, J., Zemel, R., and Culotta, A., editors, Advances in Neural Information Processing Systems, volume 23. Curran Associates, Inc.
  • van Hasselt, (2013) van Hasselt, H. (2013). Estimating the maximum expected value: An analysis of (nested) cross validation and the maximum sample average. arXiv:1302.7175.
  • Wasserkrug and Osogami, (2023) Wasserkrug, S. and Osogami, T. (2023). Who benefits from a multi-cloud market? A trading networks based analysis. arXiv:2310.12666.
  • Widmer and Leukel, (2016) Widmer, T. and Leukel, J. (2016). Efficiency of electronic service allocation with privately known quality. European Journal of Operational Research, 255(3):856–868.
  • Zhang and Zhu, (2020) Zhang, T. and Zhu, Q. (2020). Optimal two-sided market mechanism design for large-scale data sharing and trading in massive IoT networks. arXiv:1912.06229.
  • Zou, (2009) Zou, X. (2009). Double-sided auction mechanism design in electricity based on maximizing social welfare. Energy Policy, 37(11):4231–4239.

Appendix A Details

In this section, we provide full details of derivations and other details skipped in the body of the paper.

A.1 Details of Section 4

Equivalence in (8) follows from

𝔼[v(ϕ(t);tn)τn(t)tn]θ(tn)\displaystyle\mathbb{E}[v(\phi^{\star}(t);t_{n})-\tau_{n}(t)\mid t_{n}]\geq\theta(t_{n})
𝔼[m𝒩v(ϕ(t);tm)|tn]𝔼[hn(tn)|tn]θ(tn)\displaystyle\Longleftrightarrow\mathbb{E}\left[\sum_{m\in\mathcal{N}}v(\phi^{\star}(t);t_{m})\,\middle|\,t_{n}\right]-\mathbb{E}\left[h_{n}(t_{-n})\,\middle|\,t_{n}\right]\geq\theta(t_{n}) (28)
𝔼[w(t)|tn]𝔼[hn(tn)|tn]θ(tn).\displaystyle\Longleftrightarrow\mathbb{E}\left[w^{\star}(t)\,\middle|\,t_{n}\right]-\mathbb{E}\left[h_{n}(t_{-n})\,\middle|\,t_{n}\right]\geq\theta(t_{n}). (29)

Equivalence in (9) follows from

n𝒩𝔼[τn(t)]ρ\displaystyle\sum_{n\in\mathcal{N}}\mathbb{E}\left[\tau_{n}(t)\right]\geq\rho n𝒩𝔼[hn(tn)]n𝒩m𝒩n𝔼[v(ϕ(t);tm)]ρ\displaystyle\Longleftrightarrow\sum_{n\in\mathcal{N}}\mathbb{E}\left[h_{n}(t_{-n})\right]-\sum_{n\in\mathcal{N}}\sum_{m\in\mathcal{N}_{-n}}\mathbb{E}\left[v(\phi^{\star}(t);t_{m})\right]\geq\rho (30)
n𝒩𝔼[hn(tn)](N1)𝔼[w(t)]ρ,\displaystyle\Longleftrightarrow\sum_{n\in\mathcal{N}}\mathbb{E}\left[h_{n}(t_{-n})\right]-(N-1)\,\mathbb{E}\left[w^{\star}(t)\right]\geq\rho, (31)

where the last equivalence follows from the definition of w(t)w^{\star}(t) in (7).

Proposition 4.

Let the decision rule ϕ\phi^{\star} be the one that satisfies (3) and the payment rule τ=(τn)n\tau=(\tau_{n})_{n} be in the form of (6) where h=(hn)n=(hn)n=hh=(h_{n})_{n}=(h_{n}^{\star})_{n}=h^{\star} is given by the solution to the LP (10)-(12). Then the VCG mechanism (ϕ,τ)(\phi^{\star},\tau^{\star}) satisfies DSIC, DE, θ\theta-IR, and ρ\rho-WBB.

Proof.

With the equivalences (8)-(9), the constraints (11)-(12) in the LP guarantee that θ\theta-IR and ρ\rho-WBB are satisfied by feasible solutions. Since we consider the class of VCG mechanisms, DE is trivially satisfied by the definition of ϕ\phi^{\star}, and DSIC is satisfied when the payment rule is in the form of (6). Hence, all of DSIC, DE, θ\theta-IR, and ρ\rho-WBB are satisfied by (ϕ,θ)(\phi^{\star},\theta^{\star}). ∎

Algorithm 1 summarizes the protocol under the VCG mechanism discussed Section 4. In Step 3, the optimal strategy of each player is to truthfully declare its type t^n=tn\hat{t}_{n}=t_{n}. In Step 5, the LP may not be feasible, in which case the protocol may fail, or we may use another payment rule to proceed.

Algorithm 1 Protocol under the VCG mechanism
1:  Sample the type profile tt from the common prior \mathbb{P}
2:  Each player n𝒩n\in\mathcal{N} gets to know its own type tnt_{n}
3:  Each player n𝒩n\in\mathcal{N} declare their type t^n\hat{t}_{n}
4:  Determine the social decision: ϕ(t^)argmaxd𝒟n𝒩v(d;t^n)\phi^{\star}(\hat{t})\leftarrow\mathop{\rm argmax}\limits_{d\in\mathcal{D}}\displaystyle\sum_{n\in\mathcal{N}}v(d;\hat{t}_{n})
5:  hh^{\star}\leftarrow Find the optimal solution to the LP (10)-(12)
6:  Determine the payment from each player n𝒩n\in\mathcal{N} to the mediator: τn(t^)hn(t^n)m𝒩nv(ϕ(t^);t^m)\tau_{n}(\hat{t})\leftarrow h^{\star}_{n}(\hat{t}_{-n})-\displaystyle\sum_{m\in\mathcal{N}_{-n}}v(\phi^{\star}(\hat{t});\hat{t}_{m})
7:  Each player n𝒩n\in\mathcal{N} gets utility v(ϕ(t^);tn)τn(t^)v(\phi^{\star}(\hat{t});t_{n})-\tau_{n}(\hat{t})

A.2 Details of Section 5

A.2.1 Solutions that guarantee WBB

Alternatively, one may set

θ(tn)\displaystyle\theta(t_{n}) =[𝔼[w(t)tn]N1N𝔼[w(t)]]tn𝒯n,n𝒩\displaystyle=\left[\mathbb{E}[w^{\star}(t)\mid t_{n}]-\frac{N-1}{N}\mathbb{E}[w^{\star}(t)]\right]^{-}\qquad\forall t_{n}\in\mathcal{T}_{n},\forall n\in\mathcal{N} (32)
ρ\displaystyle\rho =0\displaystyle=0 (33)

to guarantee the feasibility, since

n𝒩mintn𝒯n{𝔼[wtn][𝔼[w(t)tn]N1N𝔼[w(t)]]}(N1)𝔼[w(t)]\displaystyle\sum_{n\in\mathcal{N}}\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w\mid t_{n}]-\left[\mathbb{E}[w^{\star}(t)\mid t_{n}]-\frac{N-1}{N}\,\mathbb{E}[w^{\star}(t)]\right]^{-}\right\}-(N-1)\,\mathbb{E}[w^{\star}(t)]
=n𝒩(mintn𝒯n{𝔼[wtn]N1N𝔼[w(t)][𝔼[w(t)tn]N1N𝔼[w(t)]]})\displaystyle=\sum_{n\in\mathcal{N}}\left(\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w\mid t_{n}]-\frac{N-1}{N}\,\mathbb{E}[w^{\star}(t)]-\left[\mathbb{E}[w^{\star}(t)\mid t_{n}]-\frac{N-1}{N}\,\mathbb{E}[w^{\star}(t)]\right]^{-}\right\}\right) (34)
0.\displaystyle\geq 0. (35)

In this case, player nn may incur negative utility when it has type tnt_{n} with θ(tn)<0\theta(t_{n})<0, although the loss is guaranteed to be bounded by |θ(tn)||\theta(t_{n})|.

A.2.2 On dependent types

The condition (13) may be satisfied even when types are dependent, and the optimality of our analytic solutions in Lemma 3 is guaranteed as long as (13) is satisfied. When types are dependent, however, there are cases where feasible solutions exist even when (13) is violated (Proposition 1).

In the proof of Proposition 1, we construct such a case with an extreme example of completely dependent types. However, (13) is often satisfied even in such extreme cases of completely dependent types. For example, as long as

x1x22p1px1,\displaystyle x_{1}\leq x_{2}\leq\frac{2-p}{1-p}\,x_{1}, (36)

condition (13) is satisfied in the example in the proof of Proposition 1, since then (x1,x2)(x_{1},x_{2}) satisfies (63), which corresponds to (13) in this example.

A.3 Details of Section 6

A.3.1 Upper bound

Algorithm 2 PAC-BAI(ε,δ,T)\texttt{PAC-BAI}(\varepsilon,\delta,T)
1:  Run an arbitrary (ε,δ)(\varepsilon,\delta)-PAC BAI with sample complexity TT
2:  Let I^\hat{I} be the ε\varepsilon-optimal arm identified by the algorithm
3:  Let MI^M_{\hat{I}} be the number of times the algorithm has pulled arm I^\hat{I}; if no such information is available, set MI^=0M_{\hat{I}}=0
4:  Let μ^I^\hat{\mu}_{\hat{I}} be the sample average of the MI^M_{\hat{I}} rewards obtained from arm I^\hat{I}; if MI^=0M_{\hat{I}}=0, set μ^I^=0\hat{\mu}_{\hat{I}}=0
5:  return  I^,MI^,μ^I^\hat{I},M_{\hat{I}},\hat{\mu}_{\hat{I}}
Algorithm 3 Best Mean Estimator
1:  I^,MI^,μ^I^PAC-BAI(ε,δ,T)\hat{I},M_{\hat{I}},\hat{\mu}_{\hat{I}}\leftarrow\texttt{PAC-BAI}(\varepsilon,\delta,T)
2:  if MI^<mM_{\hat{I}}<m^{\star}, where mm^{\star} is defined in Lemma 4 then
3:     Pull arm I^\hat{I} for mMI^m^{\star}-M_{\hat{I}} times
4:     Update the sample average μ^I^\hat{\mu}_{\hat{I}} of arm I^\hat{I}
5:  end if
6:  return  μ^I^\hat{\mu}_{\hat{I}}

For effective use of sample, we consider Algorithm 2, which wraps an arbitrary (ε,δ)(\varepsilon,\delta)-PAC BAI. Specifically, Algorithm 2 not only returns the arm I^\hat{I} that is identified as best by the (ε,δ)(\varepsilon,\delta)-PAC BAI but also returns the number MI^M_{\hat{I}} of i.i.d. samples that the (ε,δ)(\varepsilon,\delta)-PAC BAI has taken from I^\hat{I} as well as the corresponding sample average μI^\mu_{\hat{I}}. However, if no such information is available, it is perfectly fine, and Algorithm 2 simply returns MI^=0M_{\hat{I}}=0 and μI^=0\mu_{\hat{I}}=0 in such a case.

Notice that the naive approach of sampling each arm Θ((1/ε2)log(1/δ))\Theta((1/\varepsilon^{2})\,\log(1/\delta)) times, which also trivially falls within the upper bound in Theorem 1, would only guarantee that the best mean is estimated with the error bound of ε\varepsilon with probability at least (1δ)K(1-\delta)^{K}. Conversely, it would require sampling each arm Ω((1/ε2)log(K/δ))\Omega((1/\varepsilon^{2})\,\log(K/\delta)) times to obtain the same error bound with probability 1δ1-\delta.

A.3.2 Lower bound

The lower bound established in Lemma 4 reduces BME to the ε\varepsilon-Biased Coin Problem defined as follows:

Definition 3 (ε\varepsilon-Biased Coin Problem).

For 0<ε<10<\varepsilon<1, consider a Bernoulli random variable BB whose mean α\alpha is known to be either α+(1+ε)/2\alpha^{+}\coloneqq(1+\varepsilon)/2 or α(1ε)/2\alpha^{-}\coloneqq(1-\varepsilon)/2. The ε\varepsilon-Biased Coin Problem asks to correctly identify whether α=α+\alpha=\alpha^{+} or α=α\alpha=\alpha^{-}.

A lower bound on the sample complexity for solving the ε\varepsilon-Biased Coin Problem is known as the following lemma, for which we provide a proof in Appendix B for completeness:

Lemma 7 (Chernoff, (1972); Even-Dar et al., (2002)).

For 0<δ<1/20<\delta<1/2, any algorithm that solves the ε\varepsilon-Biased Coin Problem correctly with probability at least 1δ1-\delta has expected sample complexity at least Ω((1/ε2)log(1/δ))\Omega((1/\varepsilon^{2})\,\log(1/\delta)).

In Lemma 5, we have derived the lower bound for BME using the technique used for a lower bound for BAI in Even-Dar et al., (2002); however, as we have discussed at the end of Section 6, while our lower bound for BME is tight, the lower bound for BAI in Even-Dar et al., (2002) is not. The difference in the derived lower bound stems from the following behavior of BME and BAI when all of the arms have mean reward of α\alpha^{-} and hence are indistinguishable. The algorithm constructed in Even-Dar et al., (2002) determines that BB has mean α+\alpha^{+} when either arm i+i^{+} or arm ii^{-} is identified as the best arm. When the arms are indistinguishable, a PAC BAI would correctly identify each of the KK arms, including i+i^{+} or ii^{-}, as the best arm uniformly at random, which induces an error with probability 1/K1/K. On the other hand, the mean reward estimated by a PAC BME would be approximately correct with high probability, even when the arms are indistinguishable.

A.4 Details of Section 7

In this section, we conduct several numerical experiments to validate the effectiveness of the proposed approach and to understand its limitations. Specifically, we design our experiments to investigate the following three questions:

  1. 1.

    BME studied in Section 6 has asymptotically optimal sample complexity, but how well can we estimate the best mean when there are only a moderate number of arms?

  2. 2.

    How much can BME reduce the number of times w(t)w^{\star}(t) given by (7) is evaluated?

  3. 3.

    How well θ\theta-IR and ρ\rho-SBB are satisfied when (20) is estimated by BME rather than calculated exactly?

A.4.1 Best Mean Estimation with moderate number of arms

For BAI, existing algorithms that have asymptotically optimal sample complexity often perform poorly with moderate number of arms (Hassidim et al.,, 2020). As a result, Approximate Best Arm (Hassidim et al.,, 2020), which has optimal sample complexity, runs Naive Elimination, which has asymptotically suboptimal sample complexity, when the number of arms is below 10510^{5} or after sufficient number of suboptimal arms is eliminated. The BME algorithm studied in Section 6 has asymptotically optimal sample complexity but relies on BAI, and thus its sample complexity for moderate number of arms is not well characterized by Theorem 1. Similar to BAI, BME relies on an algorithm that performs well when there are only moderate number of arms.

Algorithm 4 Successive Elimination for Best Mean Estimation ((ε,δ)(\varepsilon,\delta)-PAC SE-BME)
0:  ε,δ\varepsilon,\delta
1:  Let {1,,K}\mathcal{R}\leftarrow\{1,\ldots,K\} be the set of remaining arms
2:  Let t0;α1t\leftarrow 0;\alpha\leftarrow 1
3:  while α>ε\alpha>\varepsilon do
4:     Pull each arm kk\in\mathcal{R} once and update the sample average μ^k\hat{\mu}_{k}
5:     tt+1t\leftarrow t+1
6:     α12tlog(π2Kt23δ)\alpha\leftarrow\sqrt{\frac{1}{2\,t}\log\left(\frac{\pi^{2}\,K\,t^{2}}{3\,\delta}\right)}
7:     for all kk\in\mathcal{R} do
8:        Remove kk from \mathcal{R} if maxμ^μ^k2α\max_{\ell\in\mathcal{R}}\hat{\mu}_{\ell}-\hat{\mu}_{k}\geq 2\,\alpha
9:     end for
10:  end while
11:  return  maxkμ^k\max_{k\in\mathcal{R}}\hat{\mu}_{k}
Algorithm 5 Successive Elimination for Best Arm Identification ((ε,δ)(\varepsilon,\delta)-PAC SE-BAI)
0:  ε,δ\varepsilon,\delta
1:  Let {1,,K}\mathcal{R}\leftarrow\{1,\ldots,K\} be the set of remaining arms
2:  Let t0;α1t\leftarrow 0;\alpha\leftarrow 1
3:  while ||>1|\mathcal{R}|>1 and α>ε2\alpha>\frac{\varepsilon}{2} do
4:     Pull each arm kk\in\mathcal{R} once and update the sample average μ^k\hat{\mu}_{k}
5:     tt+1t\leftarrow t+1
6:     α12tlog(π2Kt26δ)\alpha\leftarrow\sqrt{\frac{1}{2\,t}\log\left(\frac{\pi^{2}\,K\,t^{2}}{6\,\delta}\right)}
7:     for all kk\in\mathcal{R} do
8:        Remove kk from \mathcal{R} if maxμ^μ^k2α\max_{\ell\in\mathcal{R}}\hat{\mu}_{\ell}-\hat{\mu}_{k}\geq 2\,\alpha
9:     end for
10:  end while
11:  return  argmaxkμ^k\mathop{\rm argmax}\limits_{k\in\mathcal{R}}\hat{\mu}_{k}

In this section, we compare the performance of Successive Elimination algorithms for BAI and BME. Successive Elimination has suboptimal sample complexity, similar to Naive Elimination, but often outperforms Naive Elimination for moderate number of arms (Even-Dar et al.,, 2006). Specifically, we compare the performance of (ε,δ)(\varepsilon,\delta)-PAC Successive Elimination for BME (SE-BME; Algorithm 4) against the corresponding (ε,δ)(\varepsilon,\delta)-PAC Successive Elimination for BAI (SE-BAI; Algorithm 5). For completeness, in Appendix B, we provide standard proofs on the correctness of these algorithms, as stated in Proposition 5 and Proposition 6:

Proposition 5.

Algorithm 4 is an (ε,δ)(\varepsilon,\delta)-PAC BME.

Proposition 6.

Algorithm 5 is an (ε,δ)(\varepsilon,\delta)-PAC BAI.

There are three differences between SE-BME and SE-BAI. First, SE-BME exits the while-loop when α<ε\alpha<\varepsilon (while SE-BAI needs to wait until α<ε/2\alpha<\varepsilon/2), since it can return an overestimated mean of a suboptimal arm as long as the estimated value is within ε\varepsilon from the best mean. Second, SE-BAI exits the while-loop when only one arm remains, since it does not need to precisely estimate the mean of the identified arm. Finally, SE-BAI uses smaller αt\alpha_{t}, since it only need to consider one-sided estimation error (underestimation for best arms and overestimation for other arms).

δ=0.05\delta=0.05

Refer to caption
Refer to caption
Refer to caption

δ=0.1\delta=0.1

Refer to caption
Refer to caption
Refer to caption

δ=0.2\delta=0.2

Refer to caption

(a) ε=0.05\varepsilon=0.05

Refer to caption

(b) ε=0.1\varepsilon=0.1

Refer to caption

(c) ε=0.2\varepsilon=0.2

Figure 3: The total sample size required by (ε,δ)(\varepsilon,\delta)-PAC BME (Mean; Algorithm 4) and by (ε,δ)(\varepsilon,\delta)-PAC BAI (Arm; Algorithm 5) when arms have Bernoulli rewards with equally separated means for varying values of ε\varepsilon and δ\delta.

Figure 3 shows the total sample size required by SE-BME and by SE-BAI for varying values of ε\varepsilon and δ\delta, and for varying number KK of arms. Here, the arms have Bernoulli rewards, and their means are selected in a way they are equally separated (i.e., μk=(k0.5)/K\mu_{k}=(k-0.5)/K for k=1,,Kk=1,\ldots,K). For each datapoint, the experiments are repeated 10 times. The standard deviation of the total sample size is plotted but too small to be visible in the figure.

Overall, it can be observed that SE-BME generally requires smaller sample size than SE-BAI, except when there are only a few arms (and the means have large gaps in the setting under consideration). The efficiency of SE-BAI for a small number of arms makes intuitive sense, because the best arm can be identified without estimating the means with high accuracy.

A.4.2 Effectiveness of Best Mean Estimation in mechanism design

In this section, we quantitatively validate the effectiveness of BME in reducing the number of times w(t)w^{\star}(t) is computed when we evaluate (20). We use (ε,δ)(\varepsilon,\delta)-PAC SE-BME (Algorithm 4) as BME.

To this end, we consider the following mechanism design, motivated by the double-sided auctions for electricity (Zou,, 2009; Hobbs et al.,, 2000), where we have N=|𝒩|N=|\mathcal{N}| players, each player n𝒩n\in\mathcal{N} has KK possible types (i.e., |𝒯n|=K|\mathcal{T}_{n}|=K), with varying values of NN and KK. The KK possible types of each player are selected uniformly at random from integers, [K,K][-K,K], without replacement. We then assume, as the common prior \mathbb{P}, that the type of each player is distributed uniformly among the KK possible types and independent of the types of other players. Each player is a buyer or a seller of a single item, depending on its type. When a player nn has a positive type tnt_{n}, the player is a buyer who wants to buy a unit, whose valuation to the player is tnt_{n} (i.e., v(d;tn)=tnv(d;t_{n})=t_{n} if player nn buys a unit of the item with social decision dd). When the player has a negative type tnt_{n}, the player is a seller who wants to sell a unit, which incurs cost |tn||t_{n}| (i.e., v(d;tn)=tnv(d;t_{n})=t_{n} if player nn sells a unit of the item with social decision dd). The player does not participate in the market, when its type is zero. For a given profile of types tt, a social decision is given by a bipartite matching between buyers and sellers. In this setting, we can immediately obtain the efficient social decision (3) by greedily matching buyers of high valuations to sellers of low costs as long as the value of the buyers are higher than the costs of the sellers.

4 players

Refer to caption
Refer to caption
Refer to caption

6 players

Refer to caption
Refer to caption
Refer to caption

8 players

Refer to caption

(a) 4 types

Refer to caption

(b) 6 types

Refer to caption

(c) 8 types

Figure 4: Representative sample paths that show the estimated values of 𝔼[w(t)tn]\mathbb{E}[w^{\star}(t)\mid t_{n}] for tn𝒯nt_{n}\in\mathcal{T}_{n} against sample size used by (0.25,0.1)(0.25,0.1)-PAC SE-BME.

Figure 4 shows representative sample paths of the estimated values of 𝔼[w(t)tn]\mathbb{E}[w^{\star}(t)\mid t_{n}] for each tn𝒯nt_{n}\in\mathcal{T}_{n} when (0.25,0.1)(0.25,0.1)-PAC SE-BME111Minimization in Algorithm 4 is translated into maximization. is used to evaluate mintn𝒯n𝔼[w(t)tn]\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t_{n}] (i.e., (20) with θ(tn)=0\theta(t_{n})=0). Each panel in Figure 4 shows |𝒯n||\mathcal{T}_{n}| curves, where the red curve corresponds to the one with minimum 𝔼[w(t)tn]\mathbb{E}[w^{\star}(t)\mid t_{n}].

Observe that the types tnt_{n} (arms) that have close to the minimum mean, mintn𝒯n𝔼[w(t)tn]\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t_{n}], survive until SE-BME terminates, and their means are evaluated with sufficient accuracy. On the other hand, the types that have large means are eliminated after a relatively small number of samples without being estimated precisely, which contributes to reducing the number of evaluating the value of the efficient social decision w(t)w^{\star}(t).

Refer to caption

(a) against number of players

Refer to caption

(b) against number of types

Figure 5: The unique sample size (the number of unique tt which w(t)w^{\star}(t) is evaluated with) required by the exact computation of mintn𝒯n𝔼[w(t)tn],n𝒩\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t_{n}],\forall n\in\mathcal{N} (dashed curves) and by (0.25,0.1)(0.25,0.1)-PAC SE BME (solid curves).

In Figure 5, we study how much SE-BME can reduce the number of evaluations of w(t)w^{\star}(t) needed to estimate mintn𝒯n𝔼[w(t)t]\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t] for all n𝒩n\in\mathcal{N} with the accuracy that is shown in Figure 4 (i.e., using the same values of ε=0.25\varepsilon=0.25 and δ=0.1\delta=0.1). Recall that K=|𝒯n|,n𝒩K=|\mathcal{T}_{n}|,\forall n\in\mathcal{N} in the setting under consideration. Hence, the exact computation of mintn𝒯n𝔼[w(t)tn]\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t_{n}] for a single n𝒩n\in\mathcal{N} would require evaluating w(t)w^{\star}(t) for KNK^{N} different values of tt, but KNK^{N} evaluations of w(t)w^{\star}(t) are also sufficient to exactly compute mintn𝒯n𝔼[w(t)tn]\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t_{n}] for all n𝒩n\in\mathcal{N}, because we can cache the value of w(t)w^{\star}(t) and reuse it when it is needed. Since the computational complexity associated with evaluating w(t)w^{\star}(t) with (7) is the bottleneck, the unique sample size is what we should be interested in. Similar to exact computation, SE-BME also benefits from caching and reusing the values of w(t)w^{\star}(t). Figure 5 compares the unique sample size required by the exact computation and (0.25,0.1)(0.25,0.1)-PAC SE-BME.

Figure 5(a) implies that SE-BME (shown with solid curves) evaluates the total value of the efficient social decision, w(t)w^{\star}(t), by orders of magnitude smaller number of times than what is required by exact computation (shown with dashed curves). While the number of evaluations of w(t)w^{\star}(t) grows exponentially with the number of players (specifically, KNK^{N}) when exact computation is used, it grows only polynomially (in fact, slightly slower than linearly) when SE-BME is used. This relative insensitivity of the sample complexity of SE-BME to the number of players makes intuitive sense, because the number of players only affects the distribution of the reward and keeps the number of arms unchanged.

Figure 5(b) shows the number of evaluations of w(t)w^{\star}(t) against the number of types K=|𝒯n|K=|\mathcal{T}_{n}| for any n𝒩n\in\mathcal{N}. The advantage of SE-BME over exact computation is relatively minor when we increase the number of types instead of the number of players, since increasing the number of types directly increases the number of arms. In all cases, however, we can observe that SE-BME can significantly reduce the unique sample size.

Refer to caption

(a) against number of players

Refer to caption

(b) against number of types

Figure 6: The total sample size (the number of tt which w(t)w^{\star}(t) is evaluated with) required by the exact computation of mintn𝒯n𝔼[w(t)tn],n𝒩\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t_{n}],\forall n\in\mathcal{N} (dashed curves) and by (0.25,0.1)(0.25,0.1)-PAC SE-BME (solid curves).

Figure 6 shows the total sample size, rather than the unique sample size, required by exact computation (dashed curves) and SE-BME (solid curves). The total sample size with exact computation is NKNN\,K^{N}. While w(t)w^{\star}(t) is evaluated NN times for each tt with exact computation, SE-BME may waste evaluating the same w(t)w^{\star}(t) more often particularly when there are only a small number of players. This reduces benefits of SE-BME for total sample size, as compared to the unique sample size.

A.4.3 Individual rationality and budget balance with Best Mean Estimation

We next address the question of how well θ\theta-IR and ρ\rho-SBB are guaranteed when we estimate mintn𝒯n{𝔼[w(t)tn]θ(tn)}\min_{t_{n}\in\mathcal{T}_{n}}\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\} with BME rather than computing it exactly. We continue to use the setting of mechanism design introduced in Section A.4.2. Recall that θ\theta-IR is guaranteed when ηn\eta_{n} is given by (57) and (18), and ρ\rho-SBB is guaranteed when ηn\eta_{n} is given by (57) and (58). However, these are guaranteed only when the expected values, 𝔼[w(t)]\mathbb{E}[w^{\star}(t)] and 𝔼[w(t)tn]\mathbb{E}[w^{\star}(t)\mid t_{n}], are exactly computed. In this section, we quantitatively evaluate how well θ\theta-IR and ρ\rho-SBB are satisfied when those expected values are estimated from samples. Throughout this section, we study the case with ρ=0\rho=0 and θ0\theta\equiv 0 (i.e., θ(tn)=0,tn𝒯n,n𝒩\theta(t_{n})=0,\forall t_{n}\in\mathcal{T}_{n},\forall n\in\mathcal{N}).

θ\theta-IR is guaranteed with (57) and (18)

ρ\rho-SBB is guaranteed with (57) and (58)

Refer to caption

(a) players’ utility

Refer to caption

(b) mediator’s revenue

Refer to caption

(c) player’s utility

Refer to caption

(d) mediator’s revenue

Figure 7: The red dots show the expected utility of the players in Columns (a) and (c) and the expected revenue of the mediator in Columns (b) and (d), where analytical solutions are evaluated exactly (horizontal axes) or estimated with (0.25,0.1)(0.25,0.1)-PAC SE-BME (vertical axes) for environments with |𝒩|=8|\mathcal{N}|=8 players, each having |𝒯n|=8|\mathcal{T}_{n}|=8 possible types. The analytical solution guarantees θ\theta-IR with θ0\theta\equiv 0 in Columns (a) and (b) and ρ\rho-SBB with ρ=0\rho=0 in Columns (c) and (d). Results are plotted for 10 random seeds. Diagonal lines are also plotted to help understand where the horizontal and vertical axes are equal.

In Figure 7, we first evaluate the best mean, mintn𝒯n𝔼[w(t)t]\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t], either with exact computation or with BME, then compute ηn\eta_{n} with (57) and (18) for Columns (a)-(b) and with (57) and (58) for Columns (c)-(d), and finally evaluate the expected utility of each player (the left-hand side of (11)) for Columns (a) and (c) and the expected revenue of the mediator (the left-hand side of (12)) for Columns (b) and (d) by setting hn(tn)=ηn,tn𝒯n,n𝒩h_{n}(t_{-n})=\eta_{n},\forall t_{-n}\in\mathcal{T}_{-n},\forall n\in\mathcal{N}. Even if the best mean is estimated with BME, we evaluate the expectations on the left-hand side of (11) and (12) with exact computation, because these are the expected utility and the expected revenue that the players and the mediator will experience. Here, we repeat the experiment with 10 different random seeds, so that there are 10 data-points in Columns (b) and (d), and each of Columns (a) and (c) has 10×8×8=64010\times 8\times 8=640 data-points, where each data-point corresponds to a player of a particular type with a particular random seed.

Overall, Columns (a) and (c) of Figure 7 show that the expected utility experienced by the players is relatively insensitive to whether the best mean is evaluated with exact computation (horizontal axes) or with BME (vertical axes). Taking a closer look, we can observe that, in this particular setting, 0-IR is violated for some of the players in Column (c) even if the best mean is computed exactly, while it is guaranteed for any player of any type in Column (a) if the best mean is computed exactly.

On the other hand, as is shown in Columns (b) and (d), the mediator experiences non-negligible difference in its expected revenue depending on whether the best mean is evaluated with exact computation or with BME. It is to be expected that the mediator experiences relatively larger variance in its expected revenue, because (12) involves the summation n𝒩ηn\sum_{n\in\mathcal{N}}\eta_{n}, while (11) only involves ηn\eta_{n} for a single n𝒩n\in\mathcal{N}. In particular, it is evident in Column (d) that 0-WBB (let alone 0-SBB) is violated when the best mean is estimated with BME (vertical axes), while it is always guaranteed with exact computation (horizontal axes). In Column (b), 0-WBB is violated regardless of whether the best mean is computed exactly or with BME, since satisfying 0-WBB and 0-IR for all players (together with DE and DSIC) is impossible in this particular setting.

A simple remedy to this violation of ρ\rho-WBB is to replace the ρ\rho with a ρ>ρ\rho^{\prime}>\rho when we compute the ηn\eta_{n} from the best means, mintn𝒯n𝔼[w(t)tn]\min_{t_{n}\in\mathcal{T}_{n}}\mathbb{E}[w^{\star}(t)\mid t_{n}] for n𝒩n\in\mathcal{N}, estimated with BME. By considering the (ε,δ\varepsilon,\delta)-PAC guarantee for the error in the estimation, we can guarantee that ρ\rho-WBB is satisfied with high probability by setting an appropriate value of ρ\rho^{\prime}. Analogously, we may replace the θ(tn)\theta(t_{n}) with a θ(tn)>θ(tn)\theta^{\prime}(t_{n})>\theta(t_{n}) to provide a guaranteed that θ\theta-IR is satisfied with high probability when the best mean is estimated with BME. See Theorem 2.

θ\theta-IR is guaranteed with (57) and (18)

ρ\rho-SBB is guaranteed with (57) and (58)

Refer to caption

(a) players’ utility

Refer to caption

(b) mediator’s revenue

Refer to caption

(c) players’ utility

Refer to caption

(d) mediator’s revenue

Figure 8: The red dots show the expected utility of the players in Columns (a) and (c) and the expected revenue of the mediator in Columns (b) and (d), where analytical solutions are evaluated exactly (horizontal axes) or estimated with (0.25,0.1)(0.25,0.1)-PAC SE-BME (vertical axes) for environments with |𝒩|=8|\mathcal{N}|=8 players, each having |𝒯n|=8|\mathcal{T}_{n}|=8 possible types. The analytical solution guarantees θ\theta-IR with θ0\theta\equiv 0 in Columns (a) and (b) and ρ\rho^{\prime}-SBB with ρ=0.1\rho^{\prime}=0.1 in Columns (c) and (d). Results are plotted for 10 random seeds. Diagonal lines are also plotted to help understand where the horizontal and vertical axes are equal.

As an example, we set ρ=0.1\rho^{\prime}=0.1 in Figure 8. The consequence of replacing ρ=0\rho=0 with ρ=0.1\rho^{\prime}=0.1 is as expected. In Columns (b) and (d), the expected revenue of the mediator when the best mean is estimated with BME (Proposed approach) is shifted to the above by ρρ=0.1\rho^{\prime}-\rho=0.1. Although it may be unclear from Columns (a) and (c), the corresponding expected utility of each player is shifted to the left by (ρρ)/|𝒩|=0.0125(\rho^{\prime}-\rho)/|\mathcal{N}|=0.0125. In practice, we may choose ρ\rho^{\prime} and θ\theta^{\prime} by taking into account these shifts as well as the condition on the feasibility of LP (Corollary 1).

Refer to caption

(a) players’ utility

Refer to caption

(b) mediator’s revenue

Figure 9: Root mean squared error in (a) the expected utility of each player and (b) the expected revenue of the mediator against the total sample size, when there are |𝒩|=8|\mathcal{N}|=8 players, each with |𝒯n|=4|\mathcal{T}_{n}|=4 possible types. Here, we set δ=0.1\delta=0.1 and vary ε\varepsilon from 1.0 (purple), 0.5, 0.4, 0.3, 0.25, 0.2, to 0.15 (red).

In Figure 9, we show the Root Mean Squared Error (RMSE) in the expected utility of each player (a) and the expected revenue of the mediator (b) that are estimated with BME for the case with |𝒩|=8|\mathcal{N}|=8 players, each with |𝒯n|=4|\mathcal{T}_{n}|=4 types. Here, we fix δ=0.1\delta=0.1 and vary ε\varepsilon from 1.0 to 0.15 in the BME. For each pair of (δ,ε)(\delta,\varepsilon), the experiments are repeated 10 times with different random seeds. The total sample size increases as the value of ε\varepsilon decreases. Hence, the purple dots correspond to ε=1.0\varepsilon=1.0, and the red dots are ε=0.15\varepsilon=0.15. Overall, we can observe that RMSE can be reduced by using small ε\varepsilon at the expense of increased sample size, that relatively large values such as ε=0.5\varepsilon=0.5 gives reasonably small RMSE, and that larger values of ε\varepsilon have diminishing effects on RMSE.

A.4.4 Computational requirements

We have run all of the experiments on a single core with at most 66 GB memory without GPUs in a cloud environment. The associated source code is submitted as a supplementary material and will be open-sourced upon acceptance. Table A.4.4 summarizes the CPU time and maximum memory require to generate each figure. For example, CPU time for Figure 3(a) is the time to generate three panels in Column (a) of Figure 3. Note that the CPU time and maximum memory reported in Table A.4.4 are not optimized and include time and memory for storing intermediate results and other processing for debugging purposes; these should be understood as the computational requirements to execute the source code as is.

CPU time and maximum memory required to generate figures Figure CPU Time (seconds) Max Memory (GB) Figure 3(a) 413.1 <1<1 Figure 3(b) 81.9 <1<1 Figure 3(c) 40.4 <1<1 Figure 4(a) 0.7 <1<1 Figure 4(b) 1.1 <1<1 Figure 4(c) 80.1 1.9 Figure 5(a) and Figure 6(a) 7,883.0 65.5 Figure 5(b) and Figure 6(b) 2,935.2 17.0 Figure 7(a)-(b) and Figure 8(a)-(b) 17,531.6 1.6 Figure 7(c)-(d) and Figure 8(c)-(d) 17,416.3 1.6 Figure 9 527.9 <1<1

  • Figure 4 shows the results with one random seed, but here the CPU Time reports the average over 10 seeds, and Max Memory reports the maximum over 10 seeds.

  • Figure 8 could have been obtained by simply reusing and shifting Figure 7, but here the CPU time reports the time to generate the two figures without reuse.

A.5 Details of Section 8

A.5.1 Limitations

While the proposed approach makes major advancement in the field, it certainly has limitations. Here, we discuss four major limitations of this work as well as interesting directions of research motivated by those limitations.

First, when types are not independent between players, the sufficient condition in Lemma 1 may not be necessary (Proposition 1). This means that the LP may be feasible even when the condition in the lemma is violated, and our results do not provide optimal solutions for those cases. Further research is needed to understand exactly when the sufficient condition is also necessary. It is also important to develop efficient methods for solving the LP when the types are dependent.

Second, our mechanisms guarantee strong budget balance (SBB) and individual rationality (IR) in expectation with respect to the distribution of the players’ types, but this does not guarantee that those properties are satisfied ex post (for any realization of the types). Although the satisfaction in expectation is often sufficient for risk-neutral decision makers, it is important to let the mediator and the participants aware that they may experience negative utilities even if their expected utilities are nonnegative. It would also be an interesting direction of research to extend the proposed approach towards achieving these properties ex post.

Finally, our experiments have considered environments with up to 128 players, each with at most 16 types. Although these are substantially larger than the environment studied in Osogami et al., (2023), they certainly do not cover the scale needed for all applications. While (10-100 times) larger environments could be handled with improved implementation and greater computational resources, essentially new ideas would be needed for substantially (over 10310^{3} times) larger environments or continuous type space. It would be an interesting direction of research to identify and exploit structures of particular environments for designing scalable approaches of mechanism design for those environments.

A.5.2 Societal impacts

We expect that the proposed approach has several positive impacts on trading networks in particular and the society in general. In particular, the proposed approach enables mechanisms that can maximize the efficiency of a trading network and minimize the fees that the participants need to pay to the mediator. Also, the DSIC guaranteed by the proposed approach would make it more difficult for malicious participants to manipulate the outcome of a trading network.

On the other hand, the proposed approach might have negative impacts depending on where and how it is applied. For example, although the proposed approach guarantees individual rationality, some of the participants might get less benefits from the mechanism designed with our approach than other participants. This can happen, because maximizing the social welfare does not mean that all the participants are treated fairly. Before applying the mechanisms designed with the proposed approach, it is thus recommended to assess whether such fairness needs to be considered and to take any actions that mitigate the bias if needed.

Appendix B Proofs

B.1 Proofs on the lemmas and corollaries in Section 5

Proof of Lemma 1.

We will show that

hn(tn)\displaystyle h_{n}(t_{-n}) =ηnmintn𝒯n{𝔼[w(t)tn]θ(tn)}tn𝒯n\displaystyle=\eta_{n}\coloneqq\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\}\qquad\forall t_{-n}\in\mathcal{T}_{-n} (37)

is a feasible solution when (13) holds.

The θ\theta-IR (11) is satisfied with (37), because for any n𝒩n\in\mathcal{N} we have

𝔼[hn(tn)tn]\displaystyle\mathbb{E}[h_{n}(t_{-n})\mid t_{n}] =ηn\displaystyle=\eta_{n} (38)
=mintn𝒯n{𝔼[w(t)tn]θ(tn)}\displaystyle=\min_{t_{n}^{\prime}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t^{\prime})\mid t_{n}^{\prime}]-\theta(t_{n}^{\prime})\right\} (39)
𝔼[w(t)tn]θ(tn).\displaystyle\leq\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n}). (40)

The ρ\rho-WBB (12) is satisfied with (37), because

n𝒩𝔼[hn(tn)]\displaystyle\sum_{n\in\mathcal{N}}\mathbb{E}[h_{n}(t_{-n})] =n𝒩ηn\displaystyle=\sum_{n\in\mathcal{N}}\eta_{n} (41)
=n𝒩mintn𝒯n{𝔼[w(t)tn]θ(tn)}\displaystyle=\sum_{n\in\mathcal{N}}\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\} (42)
(N1)𝔼[w(t)]+ρ,\displaystyle\geq(N-1)\,\mathbb{E}[w^{\star}(t)]+\rho, (43)

where the inequality follows from (13). ∎

Proof of Lemma 2.

By independence, (11) is reduced to

𝔼[w(t)tn]𝔼[hn(tn)]θ(tn)tn𝒯n,n𝒩\displaystyle\mathbb{E}[w^{\star}(t)\mid t_{n}]-\mathbb{E}[h_{n}(t_{-n})]\geq\theta(t_{n})\qquad\forall t_{n}\in\mathcal{T}_{n},\forall n\in\mathcal{N} (44)
mintn𝒯n{𝔼[w(t)tn]θ(tn)}𝔼[hn(tn)]n𝒩.\displaystyle\iff\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\}\geq\mathbb{E}[h_{n}(t_{-n})]\qquad\forall n\in\mathcal{N}. (45)

This together with (12) establishes the necessity of

(N1)𝔼[w(t)]+ρ\displaystyle(N-1)\,\mathbb{E}[w^{\star}(t)]+\rho n𝒩𝔼[hn(tn)]\displaystyle\leq\sum_{n\in\mathcal{N}}\mathbb{E}[h_{n}(t_{-n})] (46)
n𝒩mintn𝒯n{𝔼[w(t)tn]θ(tn)}.\displaystyle\leq\sum_{n\in\mathcal{N}}\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\}. (47)

Proof of Lemma 3.

We first rewrite the LP (10)-(12) in the following equivalent form:

minh\displaystyle\min_{h} n𝒩tn𝒯n[tn]𝔼[hn(tn)tn]\displaystyle\qquad\sum_{n\in\mathcal{N}}\sum_{t_{n}\in\mathcal{T}_{n}}\mathbb{P}[t_{n}]\,\mathbb{E}[h_{n}(t_{-n})\mid t_{n}] (48)
s.t.\displaystyle\mathrm{s.t.} 𝔼[w(t)tn]𝔼[hn(tn)tn]θ(tn)tn𝒯n,n𝒩\displaystyle\qquad\mathbb{E}[w^{\star}(t)\mid t_{n}]-\mathbb{E}[h_{n}(t_{-n})\mid t_{n}]\geq\theta(t_{n})\qquad\forall t_{n}\in\mathcal{T}_{n},\forall n\in\mathcal{N} (49)
n𝒩tn𝒯n[tn]𝔼[hn(tn)tn](N1)𝔼[w(t)]ρ.\displaystyle\qquad\sum_{n\in\mathcal{N}}\sum_{t_{n}\in\mathcal{T}_{n}}\mathbb{P}[t_{n}]\,\mathbb{E}[h_{n}(t_{-n})\mid t_{n}]-(N-1)\,\mathbb{E}[w^{\star}(t)]\geq\rho. (50)

Then it can be easily observed that the optimal objective value must be equal to (N1)𝔼[w(t)]+ρ(N-1)\,\mathbb{E}[w^{\star}(t)]+\rho (i.e., when equality holds in (50)), since changing hh in a way it decreases the value of 𝔼[hn(tn)tn]\mathbb{E}[h_{n}(t_{-n})\mid t_{n}] only makes (49) more satisfiable222This implies that ρ\rho-SBB is satisfied whenever ρ\rho-WBB is satisfied.

Hence, to prove that any hh\in\mathcal{H} is an optimal solution, it suffices to show that (50) is satisfied with equality and (49) is satisfied with any hh\in\mathcal{H}. When hh\in\mathcal{H}, we have, for any tnt_{n}, that

𝔼[hn(tn)tn]\displaystyle\mathbb{E}[h_{n}(t_{-n})\mid t_{n}] =ηn\displaystyle=\eta_{n} (51)
mintn𝒯n{𝔼[w(t)tn]θ(tn)}\displaystyle\leq\min_{t_{n}^{\prime}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}^{\prime}]-\theta(t_{n}^{\prime})\right\} (52)
𝔼[w(t)tn]θ(tn),\displaystyle\leq\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n}), (53)

where the first inequality follows from (16). We also have

n𝒩tn𝒯n[tn]𝔼[hn(tn)tn]\displaystyle\sum_{n\in\mathcal{N}}\sum_{t_{n}\in\mathcal{T}_{n}}\mathbb{P}[t_{n}]\,\mathbb{E}[h_{n}(t_{-n})\mid t_{n}] =n𝒩tn𝒯n[tn]ηn\displaystyle=\sum_{n\in\mathcal{N}}\sum_{t_{n}\in\mathcal{T}_{n}}\mathbb{P}[t_{n}]\,\eta_{n} (54)
=n𝒩ηn\displaystyle=\sum_{n\in\mathcal{N}}\eta_{n} (55)
=(N1)𝔼[w(t)]+ρ,\displaystyle=(N-1)\,\mathbb{E}[w^{\star}(t)]+\rho, (56)

where the last equality follows from (15) and (17).

Finally, when (13) holds, \mathcal{H} is nonempty, because the following ηn\eta_{n} satisfies the conditions (15)-(17):

ηn\displaystyle\eta_{n} =mintn𝒯n{𝔼[w(t)tn]θ(tn)}δ\displaystyle=\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\}-\delta (57)
where
δ\displaystyle\delta 1N(n𝒩mintn𝒯n{𝔼[w(t)tn]θ(tn)}(N1)𝔼[w(t)]ρ).\displaystyle\coloneqq\frac{1}{N}\bigg{(}\sum_{n\in\mathcal{N}}\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\}-(N-1)\,\mathbb{E}[w^{\star}(t)]-\rho\bigg{)}. (58)

Notice that δ0\delta\geq 0 follows from (13). ∎

Proof of Proposition 1.

We construct an example that satisfies (11)-(12) but violates (13). Let 𝒩={1,2}\mathcal{N}=\{1,2\}; 𝒯n=𝒯{1,2},n𝒩\mathcal{T}_{n}=\mathcal{T}\coloneqq\{1,2\},\forall n\in\mathcal{N}; ρ=0\rho=0; θ(m)=0,m𝒯\theta(m)=0,\forall m\in\mathcal{T}. We assume that the types are completely dependent (namely, t1=t2t_{1}=t_{2} surely) and let pp be the probability that t1=t2=1t_{1}=t_{2}=1 (hence, t1=t2=2t_{1}=t_{2}=2 with probability 1p1-p).

For this example, we rewrite (11)-(12) and (13) by using xmw((m,m))x_{m}\coloneqq w^{\star}((m,m)) and ynmhn(m)y_{nm}\coloneqq h_{n}(m) for m𝒯m\in\mathcal{T} and n𝒩n\in\mathcal{N}. Notice that, for any m𝒯m\in\mathcal{T} and n𝒩n\in\mathcal{N}, we have

𝔼[w(t)tn=m]\displaystyle\mathbb{E}[w^{\star}(t)\mid t_{n}=m] =xm\displaystyle=x_{m} (59)
𝔼[hn(tn)tn=m]\displaystyle\mathbb{E}[h_{n}(t_{-n})\mid t_{n}=m] =ynm,\displaystyle=y_{nm}, (60)

since types are completely dependent. Hence, (11) is reduced to

xmynm0m𝒯,n𝒩\displaystyle x_{m}-y_{nm}\geq 0\qquad\forall m\in\mathcal{T},\forall n\in\mathcal{N} (61)

and (12) is reduced to

p(y11+y21x1)+(1p)(y12+y22x2)0.\displaystyle p\,(y_{11}+y_{21}-x_{1})+(1-p)\,(y_{12}+y_{22}-x_{2})\geq 0. (62)

On the other hand, (13) is reduced to

2min{x1,x2}px1+(1p)x2.\displaystyle 2\,\min\{x_{1},x_{2}\}\geq p\,x_{1}+(1-p)\,x_{2}. (63)

Consider the case where xm>0,m𝒯x_{m}>0,\forall m\in\mathcal{T}. In this case, (61)-(62) suggest that (11)-(12) are satisfied as long as ynmy_{nm} satisfies

xm2ynmxmm𝒯,n𝒩,\displaystyle\frac{x_{m}}{2}\leq y_{nm}\leq x_{m}\qquad\forall m\in\mathcal{T},\forall n\in\mathcal{N}, (64)

whether (63) is satisfied or not. Indeed, (64) can be met even if (63) is violated, for example when p=12,x1=1,x2=4,ynm=23xm,m𝒯,n𝒩p=\frac{1}{2},x_{1}=1,x_{2}=4,y_{nm}=\frac{2}{3}x_{m},\forall m\in\mathcal{T},\forall n\in\mathcal{N}; this serves as an desired example, concluding the proof. ∎

Proof of Corollary 2.

The sufficiency follows from Lemma 3. The necessity follows in exactly the same way as the proof of Lemma 2. ∎

Proof of Corollary 3.

By (15), we have

n𝒩ηn(N1)𝔼[w(t)]ρ\displaystyle\sum_{n\in\mathcal{N}}\eta_{n}-(N-1)\,\mathbb{E}[w^{\star}(t)]-\rho
=n𝒩mintn𝒯n{𝔼[w(t)tn]θ(tn)}n𝒩δn(N1)𝔼[w(t)]ρ\displaystyle=\sum_{n\in\mathcal{N}}\min_{t_{n}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})\right\}-\sum_{n\in\mathcal{N}}\delta_{n}-(N-1)\,\mathbb{E}[w^{\star}(t)]-\rho (65)
=0,\displaystyle=0, (66)

where the last equality follows from (17). Hence, (12) is satisfied with equality. ∎

Proof of Corollary 4.

This corollary can be proved analogously to Corollary 3. ∎

Proof of Corollary 5.

By (15), for any tn𝒯nt_{n}\in\mathcal{T}_{n} and n𝒩n\in\mathcal{N}, we have

𝔼[w(t)tn]ηnθ(tn)\displaystyle\mathbb{E}[w^{\star}(t)\mid t_{n}]-\eta_{n}-\theta(t_{n})
=𝔼[w(t)tn]θ(tn)(mintn𝒯n{𝔼[w(t)tn]θ(tn)})+δn,\displaystyle=\mathbb{E}[w^{\star}(t)\mid t_{n}]-\theta(t_{n})-\left(\min_{t_{n}^{\prime}\in\mathcal{T}_{n}}\left\{\mathbb{E}[w^{\star}(t)\mid t_{n}^{\prime}]-\theta(t_{n}^{\prime})\right\}\right)+\delta_{n}, (67)

which is nonnegative by (16), and hence (11) holds. ∎

B.2 Proofs of the lemmas, theorem, and proposition in Section 6

Proof of Lemma 4.

Since the sample complexity of PAC-BAI in Step 1 is MM and Step 3 pulls an arm at most mm^{\star} times, the sample complexity of Algorithm 3 is at most M+mM+m^{\star}. Hence, it remains to prove (23).

Recall that I^\hat{I} is a random variable representing the index of the best arm returned by Algorithm 2 (PAC-BAI). Then we have the following bound:

Pr(|μ^I^μ|>32ε)\displaystyle\Pr\left(|\hat{\mu}_{\hat{I}}-\mu_{\star}|>\frac{3}{2}\varepsilon\right)
=Pr(μ^I^>μ+32ε)+Pr(μ^I^<μ32ε)\displaystyle=\Pr\left(\hat{\mu}_{\hat{I}}>\mu_{\star}+\frac{3}{2}\varepsilon\right)+\Pr\left(\hat{\mu}_{\hat{I}}<\mu_{\star}-\frac{3}{2}\varepsilon\right) (68)
Pr(μ^I^>μI^+32ε)\displaystyle\leq\Pr\left(\hat{\mu}_{\hat{I}}>\mu_{\hat{I}}+\frac{3}{2}\varepsilon\right)
+Pr({μ^I^<μ32ε}{μI^<με})+Pr({μ^I^<μ32ε}{μI^με})\displaystyle\hskip 8.53581pt+\Pr\left(\left\{\hat{\mu}_{\hat{I}}<\mu_{\star}-\frac{3}{2}\varepsilon\right\}\cap\left\{\mu_{\hat{I}}<\mu_{\star}-\varepsilon\right\}\right)+\Pr\left(\left\{\hat{\mu}_{\hat{I}}<\mu_{\star}-\frac{3}{2}\varepsilon\right\}\cap\left\{\mu_{\hat{I}}\geq\mu_{\star}-\varepsilon\right\}\right) (69)
Pr(μ^I^>μI^+32ε)+Pr(μI^<με)+Pr({μ^I^<μ32ε}{μI^με})\displaystyle\leq\Pr\left(\hat{\mu}_{\hat{I}}>\mu_{\hat{I}}+\frac{3}{2}\varepsilon\right)+\Pr\left(\mu_{\hat{I}}<\mu_{\star}-\varepsilon\right)+\Pr\left(\left\{\hat{\mu}_{\hat{I}}<\mu_{\star}-\frac{3}{2}\varepsilon\right\}\cap\left\{\mu_{\hat{I}}\geq\mu_{\star}-\varepsilon\right\}\right) (70)
Pr(μ^I^>μI^+32ε)+Pr(μI^<με)+Pr(μ^I^<μI^12ε)\displaystyle\leq\Pr\left(\hat{\mu}_{\hat{I}}>\mu_{\hat{I}}+\frac{3}{2}\varepsilon\right)+\Pr\left(\mu_{\hat{I}}<\mu_{\star}-\varepsilon\right)+\Pr\left(\hat{\mu}_{\hat{I}}<\mu_{\hat{I}}-\frac{1}{2}\varepsilon\right) (71)
Pr(μ^I^>μI^+32ε)+δ+Pr(μ^I^<μI^12ε),\displaystyle\leq\Pr\left(\hat{\mu}_{\hat{I}}>\mu_{\hat{I}}+\frac{3}{2}\varepsilon\right)+\delta+\Pr\left(\hat{\mu}_{\hat{I}}<\mu_{\hat{I}}-\frac{1}{2}\varepsilon\right), (72)

where the last inequality follows from PAC(ε,δ\varepsilon,\delta) of BAI.

Since μ^I^\hat{\mu}_{\hat{I}} is the average of Mmax{MI^,m}M^{\prime}\coloneqq\max\{M_{\hat{I}},m^{\star}\} samples from arm I^\hat{I}, where MI^M_{\hat{I}} is the number of times Algorithm 2 (PAC-BAI) has pulled arm I^\hat{I}, we have MmM^{\prime}\geq m^{\star}. Hence, applying Hoeffding’s inequality to the last term of (72), we obtain

Pr(μ^I^<μI^12ε)\displaystyle\Pr\left(\hat{\mu}_{\hat{I}}<\mu_{\hat{I}}-\frac{1}{2}\varepsilon\right) =nm,k[1,K]Pr(μ^k<μk12ε|M=n,I^=k)Pr(M=n,I^=k)\displaystyle=\sum_{n\geq m^{\star},k\in[1,K]}\Pr\left(\hat{\mu}_{k}<\mu_{k}-\frac{1}{2}\varepsilon\,\middle|\,M^{\prime}=n,\hat{I}=k\right)\,\Pr(M^{\prime}=n,\hat{I}=k) (73)
nm,k[1,K]exp(2(12ε)2n)Pr(M=n,I^=k)\displaystyle\leq\sum_{n\geq m^{\star},k\in[1,K]}\exp\left(-2\left(\frac{1}{2}\varepsilon\right)^{2}n\right)\,\Pr(M^{\prime}=n,\hat{I}=k) (74)
exp(2(12ε)2m),\displaystyle\leq\exp\left(-2\left(\frac{1}{2}\varepsilon\right)^{2}m^{\star}\right), (75)
where the first inequality is obtained by applying Hoeffding’s inequality to the sample mean μ^k\hat{\mu}_{k} of nn independent random variables having support in [0,1][0,1], and the second inequality follows from nmn\geq m^{\star}. We can also show the following inequality in an analogous manner:
Pr(μ^I^>μI^+32ε)\displaystyle\Pr\left(\hat{\mu}_{\hat{I}}>\mu_{\hat{I}}+\frac{3}{2}\varepsilon\right) exp(2(32ε)2m).\displaystyle\leq\exp\left(-2\left(\frac{3}{2}\varepsilon\right)^{2}m^{\star}\right). (76)

By applying (75)-(76) to (72), we finally establish the bound to be shown:

Pr(|μ^I^μ|>32ε)\displaystyle\Pr\left(|\hat{\mu}_{\hat{I}}-\mu_{\star}|>\frac{3}{2}\varepsilon\right) δ+exp(2(32ε)2m)+exp(2(12ε)2m)\displaystyle\leq\delta+\exp\left(-2\,\left(\frac{3}{2}\varepsilon\right)^{2}\,m^{\star}\right)+\exp\left(-2\,\left(\frac{1}{2}\varepsilon\right)^{2}\,m^{\star}\right) (77)
δ+exp(2(32ε)22ε2log1.22δ)+exp(2(12ε)22ε2log1.22δ)\displaystyle\leq\delta+\exp\left(-2\,\left(\frac{3}{2}\varepsilon\right)^{2}\,\frac{2}{\varepsilon^{2}}\log\frac{1.22}{\delta}\right)+\exp\left(-2\,\left(\frac{1}{2}\varepsilon\right)^{2}\,\frac{2}{\varepsilon^{2}}\log\frac{1.22}{\delta}\right)
by the definition of mm^{\star} (78)
=δ+(δ1.22)9+δ1.22\displaystyle=\delta+\left(\frac{\delta}{1.22}\right)^{9}+\frac{\delta}{1.22} (79)
(1+11.229+11.22)δ\displaystyle\leq\left(1+\frac{1}{1.22^{9}}+\frac{1}{1.22}\right)\delta (80)
2δ.\displaystyle\leq 2\,\delta. (81)

Proof of Lemma 7.

Although the lemma is stated in Even-Dar et al., (2002) with reference to Chernoff, (1972), this specific lemma is neither stated nor proved explicitly in Chernoff, (1972). For completeness, here, we prove the lemma following the general methodology provided in Chernoff, (1972). Specifically, we derive the expected sample size required by the sequential probability-ratio test (SPRT; Section 10 of Chernoff, (1972)), whose optimality (Theorem 12.1 of Chernoff, (1972)) will then establish the lemma.

Consider two hypotheses, θ1\theta_{1} and θ2\theta_{2}, for the probability distribution \mathbb{P} of a random variable XX, which takes either the value of 1 or 1-1, where

(X=1θ1)\displaystyle\mathbb{P}(X=1\mid\theta_{1}) =1+ε2\displaystyle=\frac{1+\varepsilon}{2} (82)
(X=1θ2)\displaystyle\mathbb{P}(X=1\mid\theta_{2}) =1ε2\displaystyle=\frac{1-\varepsilon}{2} (83)

Consider the SPRT procedure that takes i.i.d. samples, X1,X2,,XNX_{1},X_{2},\ldots,X_{N}, from \mathbb{P} until the stopping time NN when

λN\displaystyle\lambda_{N} n=1N(Xnθ1)(Xnθ2)=n=1N(ε+1ε1)Xn\displaystyle\coloneqq\prod_{n=1}^{N}\frac{\mathbb{P}(X_{n}\mid\theta_{1})}{\mathbb{P}(X_{n}\mid\theta_{2})}=\prod_{n=1}^{N}\left(\frac{\varepsilon+1}{\varepsilon-1}\right)^{X_{n}} (84)

hits either AA\in\mathbb{R} or 1/A1/A. When λN\lambda_{N} hits AA, we identify θ1\theta_{1} as the correct hypothesis. When λN\lambda_{N} hits 1/A1/A, we identify θ2\theta_{2} as the correct hypothesis.

Let

SNlogλN=n=1NXnlog1+ε1ε.\displaystyle S_{N}\coloneqq\log\lambda_{N}=\sum_{n=1}^{N}X_{n}\,\log\frac{1+\varepsilon}{1-\varepsilon}. (85)

Since NN is a stopping time, by Wald’s lemma, we have

𝔼[SNθ1]\displaystyle\mathbb{E}[S_{N}\mid\theta_{1}] =𝔼[Nθ1]𝔼[Xθ1]log1+ε1ε\displaystyle=\mathbb{E}[N\mid\theta_{1}]\,\mathbb{E}[X\mid\theta_{1}]\,\log\frac{1+\varepsilon}{1-\varepsilon} (86)
=𝔼[Nθ1]εlog1+ε1ε\displaystyle=\mathbb{E}[N\mid\theta_{1}]\,\varepsilon\,\log\frac{1+\varepsilon}{1-\varepsilon} (87)
𝔼[SNθ2]\displaystyle\mathbb{E}[S_{N}\mid\theta_{2}] =𝔼[Nθ2]εlog1+ε1ε.\displaystyle=-\mathbb{E}[N\mid\theta_{2}]\,\varepsilon\,\log\frac{1+\varepsilon}{1-\varepsilon}. (88)

Let δ\delta be the probability of making the error in identifying the correct hypothesis. Then we must have

𝔼[SNθ1]\displaystyle\mathbb{E}[S_{N}\mid\theta_{1}] =(1δ)logA+δlog(1/A)\displaystyle=(1-\delta)\,\log A+\delta\,\log(1/A) (89)
𝔼[SNθ2]\displaystyle\mathbb{E}[S_{N}\mid\theta_{2}] =δlogA+(1δ)log(1/A).\displaystyle=\delta\,\log A+(1-\delta)\,\log(1/A). (90)

By (86)-(90), we have

𝔼[Nθ1]=𝔼[Nθ2]=12δεlog1+ε1εlogA.\displaystyle\mathbb{E}[N\mid\theta_{1}]=\mathbb{E}[N\mid\theta_{2}]=\frac{1-2\,\delta}{\varepsilon\log\frac{1+\varepsilon}{1-\varepsilon}}\,\log A. (91)

Now, notice that SNS_{N} hits logA\log A when we have

|{n:Xn=1}||{n:Xn=1}|logAlog1+ε1ε\displaystyle|\{n:X_{n}=1\}|-|\{n:X_{n}=-1\}|\geq\frac{\log A}{\log\frac{1+\varepsilon}{1-\varepsilon}} (92)

for the first time and hits logA-\log A when we have

|{n:Xn=1}||{n:Xn=1}|logAlog1+ε1ε\displaystyle|\{n:X_{n}=-1\}|-|\{n:X_{n}=1\}|\geq\frac{\log A}{\log\frac{1+\varepsilon}{1-\varepsilon}} (93)

for the first time. Hence, by the gambler’s ruin probability, we have

δ\displaystyle\delta =1(1+ε1ε)logAlog1+ε1ε1(1+ε1ε)2logAlog1+ε1ε=11+(1+ε1ε)logAlog1+ε1ε,\displaystyle=\frac{1-\left(\frac{1+\varepsilon}{1-\varepsilon}\right)^{\frac{\log A}{\log\frac{1+\varepsilon}{1-\varepsilon}}}}{1-\left(\frac{1+\varepsilon}{1-\varepsilon}\right)^{2\frac{\log A}{\log\frac{1+\varepsilon}{1-\varepsilon}}}}=\frac{1}{1+\left(\frac{1+\varepsilon}{1-\varepsilon}\right)^{\frac{\log A}{\log\frac{1+\varepsilon}{1-\varepsilon}}}}, (94)

which implies

A\displaystyle A =1δδ.\displaystyle=\frac{1-\delta}{\delta}. (95)

Plugging the last expression into (91), we obtain

𝔼[Nθ1]=𝔼[Nθ2]=12δεlog1+ε1εlog1δδ=O(1ε2log1δ).\displaystyle\mathbb{E}[N\mid\theta_{1}]=\mathbb{E}[N\mid\theta_{2}]=\frac{1-2\,\delta}{\varepsilon\log\frac{1+\varepsilon}{1-\varepsilon}}\,\log\frac{1-\delta}{\delta}=O\left(\frac{1}{\varepsilon^{2}}\log\frac{1}{\delta}\right). (96)

Proof of Lemma 5.

We will construct an algorithm that correctly identifies the mean α\alpha of BB with probability at least δ\delta using at most M/KM/K samples of BB in expectation given the access to an (ε/2,δ/2)(\varepsilon/2,\delta/2)-PAC BME with sample complexity MM. The algorithm first draws i+i^{+} and ii^{-} independently from a uniformly distribution over [1,K][1,K].

Consider two environments of KK-armed bandit, +\mathcal{E}^{+} and \mathcal{E}^{-}, where every arm gives the reward according to the Bernoulli distribution with mean α=(1ε)/2\alpha^{-}=(1-\varepsilon)/2 except that the reward of arm i+i^{+} in +\mathcal{E}^{+} has the same distribution as BB and the reward of arm ii^{-} in \mathcal{E}^{-} has the same distribution as 1B1-B. Note that the algorithm can simulate the reward with the known mean α\alpha^{-} using the algorithm’s internal random number generators. Only when the algorithm pulls arm i+i^{+} of +\mathcal{E}^{+} or arm ii^{-} of \mathcal{E}^{-}, it uses the sample of BB, which contributes to the sample complexity of the algorithm.

The algorithm then runs two copies of the (ε/2,δ/2)(\varepsilon/2,\delta/2)-PAC BME with sample complexity MM in parallel: one referred to as BME+ is run on 𝒯+\mathcal{T}^{+}, and the other referred to as BME- is run on 𝒯\mathcal{T}^{-}. At each step, let M+M^{+} be the number of samples BME+ has taken from arm i+i^{+} by that step and MM^{-} be the corresponding number BME- has taken from arm ii^{-}. If M+<MM^{+}<M^{-}, the algorithm lets BME+ pull an arm; otherwise, the algorithm lets BME- pull an arm. Therefore, |M+M|1|M^{+}-M^{-}|\leq 1 at any step.

This process is continued until one of BME+ and BME- terminates and returns an estimate μ^\hat{\mu} of the best mean. If BME+ terminates first, then the algorithm determines that α=α\alpha=\alpha^{-} if μ^<1/2\hat{\mu}<1/2 and that α=α+\alpha=\alpha^{+} otherwise. If BME- terminates first, then the algorithm determines that α=α+\alpha=\alpha^{+} if μ^<1/2\hat{\mu}<1/2 and that α=α\alpha=\alpha^{-} otherwise. Due to the (ε/2,δ/2)(\varepsilon/2,\delta/2)-PAC property of BME+ and BME-, the algorithm correctly identifies the mean of BB with probability at least 1δ1-\delta. Formally, if α=α+\alpha=\alpha^{+}, we have

Pr(μ^<12)\displaystyle\Pr\left(\hat{\mu}<\frac{1}{2}\right) =Pr(μ^<12|BME+ terminates first)Pr(BME+ terminates first)\displaystyle=\Pr\left(\hat{\mu}<\frac{1}{2}\,\middle|\,\mbox{BME}^{+}\mbox{ terminates first}\right)\,\Pr(\mbox{BME}^{+}\mbox{ terminates first})
+Pr(μ^<12|BME terminates first)Pr(BME terminates first)\displaystyle\hskip 8.53581pt+\Pr\left(\hat{\mu}<\frac{1}{2}\,\middle|\,\mbox{BME}^{-}\mbox{ terminates first}\right)\,\Pr(\mbox{BME}^{-}\mbox{ terminates first}) (97)
δ2+δ2\displaystyle\leq\frac{\delta}{2}+\frac{\delta}{2} (98)
=δ.\displaystyle=\delta. (99)

Analogously, Pr(μ^<12)δ\Pr\left(\hat{\mu}<\frac{1}{2}\right)\leq\delta can be shown if α=α\alpha=\alpha^{-}.

What remains to prove is the sample complexity of the algorithm. Recall that each of BME+ and BME- pulls arms at most MM times before it terminates due to their sample complexity. Notice that the arms in \mathcal{E}^{-} are indistinguishable when α=α\alpha=\alpha^{-}, and the arms in +\mathcal{E}^{+} are indistinguishable when α=α+\alpha=\alpha^{+}. Therefore, at least one of BME+ and BME- is run on the environment where the arms are indistinguishable. Since ii^{-} and ii^{-} are sampled uniformly at random from [1,K][1,K], BME (either BME+ or BME-) would take at most M/nM/n samples from BB in expectation if the arms are indistinguishable. Since |M+M|1|M^{+}-M^{-}|\leq 1, we establish that the sample complexity of the algorithm is O(M/n)O(M/n) in expectation. ∎

Proof of Lemma 6.

Note that the PAC estimators can give independent estimates, κ~n(θ)\tilde{\kappa}_{n}(\theta) for n𝒩n\in\mathcal{N} and λ~\tilde{\lambda}, such that

Pr(|κ~n(θ)κn(θ)|ε)\displaystyle\Pr\left(|\tilde{\kappa}_{n}(\theta)-\kappa_{n}(\theta)|\leq\varepsilon^{\prime}\right) 1δn𝒩\displaystyle\geq 1-\delta^{\prime}\qquad\forall n\in\mathcal{N} (100)
Pr(|λ~(θ)λ(θ)|ε′′)\displaystyle\Pr\left(|\tilde{\lambda}(\theta)-\lambda(\theta)|\leq\varepsilon^{\prime\prime}\right) 1δ,\displaystyle\geq 1-\delta^{\prime}, (101)

which imply

Pr(|λ~(θ)λ(θ)|ε′′,|κ~n(θ)κn(θ)|ε,n𝒩)\displaystyle\Pr\left(|\tilde{\lambda}(\theta)-\lambda(\theta)|\leq\varepsilon^{\prime\prime},|\tilde{\kappa}_{n}(\theta)-\kappa_{n}(\theta)|\leq\varepsilon^{\prime},\forall n\in\mathcal{N}\right) (1δ)N+1.\displaystyle\geq(1-\delta^{\prime})^{N+1}. (102)

Hence, with probability at least (1δ)N+1(1-\delta^{\prime})^{N+1}, the expected utility of player nn given its type tnt_{n} (recall (8)) is

𝔼[w(t)tn](κ~n(θ)d~n)\displaystyle\mathop{\mathbb{E}}\limits[w^{\star}(t)\mid t_{n}]-(\tilde{\kappa}_{n}(\theta)-\tilde{d}_{n}) 𝔼[w(t)tn]κn(θ)ε+ε′′′\displaystyle\geq\mathop{\mathbb{E}}\limits[w^{\star}(t)\mid t_{n}]-\kappa_{n}(\theta)-\varepsilon^{\prime}+\varepsilon^{\prime\prime\prime} (103)
θ(tn)(εε′′′),\displaystyle\geq\theta(t_{n})-(\varepsilon^{\prime}-\varepsilon^{\prime\prime\prime}), (104)

where the first inequality follows from the PAC bound on κ~n(θ)\tilde{\kappa}_{n}(\theta) and the definition of d~n\tilde{d}_{n}, and the last inequality follows from the original guarantee when κn(θ)\kappa_{n}(\theta) is exactly computed, and the expected revenue of the mediator (recall (9)) is

n𝒩(κ~n(θ)d~n)(N1)𝔼[w(t)]\displaystyle\sum_{n\in\mathcal{N}}(\tilde{\kappa}_{n}(\theta)-\tilde{d}_{n})-(N-1)\,\mathop{\mathbb{E}}\limits[w^{\star}(t)] =(N1)(λ~+ε′′′)(N1)𝔼[w(t)]\displaystyle=(N-1)\,(\tilde{\lambda}+\varepsilon^{\prime\prime\prime})-(N-1)\,\mathop{\mathbb{E}}\limits[w^{\star}(t)] (105)
ρ+(N1)(ε′′′ε′′),\displaystyle\geq\rho+(N-1)\,(\varepsilon^{\prime\prime\prime}-\varepsilon^{\prime\prime}), (106)

where the equality follows from the definition of d~\tilde{d}, and the inequality follows from the PAC bound on λ~\tilde{\lambda}. ∎

Proof of Theorem 2.

With the choice of the parameters in the theorem, we have

θ(εε′′′)\displaystyle\theta-(\varepsilon^{\prime}-\varepsilon^{\prime\prime\prime}) =θ\displaystyle=\theta (107)
(ρ(N1)(ε′′ε′′′′))\displaystyle(\rho-(N-1)\,(\varepsilon^{\prime\prime}-\varepsilon^{\prime\prime\prime\prime})) =ρ\displaystyle=\rho (108)
(1δ)1/(N+1)\displaystyle(1-\delta^{\prime})^{1/(N+1)} =1δ.\displaystyle=1-\delta. (109)

Hence, Lemma 6 guarantees that the constant pivot rule hn(tn)=κ~n(θ)d~nh_{n}(t_{-n})=\tilde{\kappa}_{n}(\theta)-\tilde{d}_{n} satisfies DSIC, DE, θ\theta-IR, and ρ\rho-WBB with probability 1δ1-\delta. ∎

Proof of Proposition 3.

The constant pivot rule in Theorem 2 can be learned with NN independent runs of an (ε,δ)(\varepsilon,\delta^{\prime})-PAC BME and a single run of an (ε,δ)(\varepsilon,\delta^{\prime})-PAC estimator for an expectation, whose overall sample complexity is O(N(K/ε2)log(1/δ))O(N\,(K/\varepsilon^{2})\,\log(1/\delta^{\prime})). The proposition can then be established by substituting δ=1(1δ)1/(N+1)\delta^{\prime}=1-(1-\delta)^{1/(N+1)}. ∎

B.3 Proof of the propositions in Section 7

Proof of Proposition 5.

Let μmaxk[1,K]μk\mu_{\star}\coloneqq\max_{k\in[1,K]}\mu_{k} be the best mean, μ^\hat{\mu} be the best mean estimated by Algorithm 4, and μ^k(t)\hat{\mu}_{k}^{(t)} be the average of the first tt samples from arm kk. Let αt12tlog(π2Kt23δ)\alpha_{t}\coloneqq\sqrt{\frac{1}{2\,t}\log\left(\frac{\pi^{2}\,K\,t^{2}}{3\,\delta}\right)}. Then we have

Pr(|μ^μ|ε)\displaystyle\Pr(|\hat{\mu}-\mu_{\star}|\leq\varepsilon)
Pr(At every iteration, the error in the estimated mean is less than αt for any arm in .333segevw@il.ibm.com)\displaystyle\geq\Pr(\mbox{At every iteration, the error in the estimated mean is less than $\alpha_{t}$ for any arm in $\mathcal{R}$.}) (110)
=Pr(t=1k[1,K]{|{μ^k(t)μk|<αt})\displaystyle=\Pr\left(\bigcap_{t=1}^{\infty}\bigcap_{k\in[1,K]}\{|\{\hat{\mu}_{k}^{(t)}-\mu_{k}|<\alpha_{t}\}\right) (111)
=1Pr(t=1k[1,K]{|{μ^k(t)μk|αt})\displaystyle=1-\Pr\left(\bigcup_{t=1}^{\infty}\bigcup_{k\in[1,K]}\{|\{\hat{\mu}_{k}^{(t)}-\mu_{k}|\geq\alpha_{t}\}\right) (112)
1t=1k[1,K]Pr({|{μ^k(t)μk|αt})by union bound\displaystyle\geq 1-\sum_{t=1}^{\infty}\sum_{k\in[1,K]}\Pr\left(\{|\{\hat{\mu}_{k}^{(t)}-\mu_{k}|\geq\alpha_{t}\}\right)\quad\mbox{by union bound} (113)
12Kt=1exp(2αt2t)by Hoeffding’s inequality\displaystyle\geq 1-2\,K\,\sum_{t=1}^{\infty}\exp\left(-2\,\alpha_{t}^{2}\,t\right)\quad\mbox{by Hoeffding's inequality} (114)
=1δ6π2t=11t2by definition of αt\displaystyle=1-\delta\,\frac{6}{\pi^{2}}\,\sum_{t=1}^{\infty}\frac{1}{t^{2}}\quad\mbox{by definition of $\alpha_{t}$} (115)
=1δ.\displaystyle=1-\delta. (116)
33footnotetext: This condition suffices because it ensures that the best arms always remain in \mathcal{R}.

Proof of Proposition 6.

Let αtlog(π2Kt26δ)/(2t)\alpha_{t}\coloneqq\sqrt{\log\left(\frac{\pi^{2}\,K\,t^{2}}{6\,\delta}\right)/(2\,t)}. Let \mathcal{B} be the set of the (strictly) best arms. Let μ^k(t)\hat{\mu}_{k}^{(t)} be the average of the first tt samples from arm kk. Then we have

Pr(Algorithm 5 selects an ε-best arm.)\displaystyle\Pr(\mbox{Algorithm~{}\ref{alg:SE-BAI} selects an $\varepsilon$-best arm.})
Pr(At every iteration, all arms in  remain in  and any arm in  is 2αt-best.444This condition sufficies because the algorithm stops either when |R|=1 or when αtε/2, which implies that only ε-best arms are in R when the algorithm stops.)\displaystyle\geq\Pr(\mbox{At every iteration, all arms in $\mathcal{B}$ remain in $\mathcal{R}$ and any arm in $\mathcal{R}$ is $2\alpha_{t}$-best.}) (117)
Pr(t=1k{μ^k(t)>μkαt}{μ^(t)<μ+αt})\displaystyle\geq\Pr\left(\bigcap_{t=1}^{\infty}\bigcap_{k\in\mathcal{B}}\{\hat{\mu}_{k}^{(t)}>\mu_{k}-\alpha_{t}\}\bigcap_{\ell\not\in\mathcal{B}}\{\hat{\mu}_{\ell}^{(t)}<\mu_{\ell}+\alpha_{t}\}\right) (118)
=1Pr(i=1k{μ^k(t)μkαt}{μ^(t)μ+αt})\displaystyle=1-\Pr\left(\bigcup_{i=1}^{\infty}\bigcup_{k\in\mathcal{B}}\{\hat{\mu}_{k}^{(t)}\leq\mu_{k}-\alpha_{t}\}\bigcup_{\ell\not\in\mathcal{B}}\{\hat{\mu}_{\ell}^{(t)}\geq\mu_{\ell}+\alpha_{t}\}\right) (119)
1t=1(kPr(μ^k(t)μkαt)+Pr(μ^(t)μ+αt))by union bound\displaystyle\geq 1-\sum_{t=1}^{\infty}\left(\sum_{k\in\mathcal{B}}\Pr(\hat{\mu}_{k}^{(t)}\leq\mu_{k}-\alpha_{t})+\sum_{\ell\not\in\mathcal{B}}\Pr(\hat{\mu}_{\ell}^{(t)}\geq\mu_{\ell}+\alpha_{t})\right)\quad\mbox{by union bound} (120)
1t=1Kexp(2αt2t)by Hoeffding’s inequality\displaystyle\geq 1-\sum_{t=1}^{\infty}K\,\exp(-2\,\alpha_{t}^{2}\,t)\quad\mbox{by Hoeffding's inequality} (121)
=1δ6π2t=11t2by the definition of αt\displaystyle=1-\delta\,\frac{6}{\pi^{2}}\,\sum_{t=1}^{\infty}\frac{1}{t^{2}}\quad\mbox{by the definition of $\alpha_{t}$} (122)
=1δ.\displaystyle=1-\delta. (123)