This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Barriers to Welfare Maximization with
No-Regret Learning

Ioannis Anagnostides Alkis Kalavasis Yale University Tuomas Sandholm Optimized Markets, Inc.
Abstract

A celebrated result in the interface of online learning and game theory guarantees that the repeated interaction of no-regret players leads to a coarse correlated equilibrium (CCE)—a natural game-theoretic solution concept. Despite the rich history of this foundational problem and the tremendous interest it has received in recent years, a basic question still remains open: how many iterations are needed for no-regret players to approximate an equilibrium? In this paper, we establish the first computational lower bounds for that problem in two-player (general-sum) games under the constraint that the CCE reached approximates the optimal social welfare (or some other natural objective). From a technical standpoint, our approach revolves around proving lower bounds for computing a near-optimal TT-sparse CCE—a mixture of TT product distributions, thereby circumscribing the iteration complexity of no-regret learning even in the centralized model of computation. Our proof proceeds by extending a classical reduction of Gilboa and Zemel [1989] for optimal Nash to sparse (approximate) CCE. In particular, we show that the inapproximability of maximum clique precludes attaining any non-trivial sparsity in polynomial time. Moreover, we strengthen our hardness results to apply in the low-precision regime as well via the planted clique conjecture.

1 Introduction

One of the most influential results in the interface of algorithmic game theory and online learning is the realization that repeated play under no-regret—a basic notion of hindsight rationality— leads to a natural game-theoretic solution concept known as coarse correlated equilibrium (CCE) [HM00, FV97]. Many ubiquitous algorithms guarantee the no-regret property, including (online) gradient descent and multiplicative weights update, and so one should expect CCE to arise from the repeated interaction of rational players—as it has been corroborated empirically [NST15, KN22]. From an algorithmic standpoint, perhaps the most well-studied question that emerged from that connection concerns the number of iterations needed to approximate an equilibrium. Remarkably, although this problem traces back to the early pioneering works of [Bla56] and  [Rob51] in the 1950s, it remains poorly understood. This stands in contrast to the so-called adversarial regime, wherein a learner engages repeatedly with an adversarial environment acting so as to maximize the player’s regret; in that setting, the minimax regret of the learner has long been resolved in the online learning literature [LW94]. Yet, it turns out that substantially improved guarantees are possible when the learner is instead competing against other learning players, as witnesses by a flurry of recent results (e.g.[DDK15, Syr+15, RS13, DFG21, PSS22]). Indeed, such learning dynamics have emerged as a key component in practical equilibrium computation [BS19a, Bow+15, Bak+22, Per+22], proving to be more scalable than traditional linear programming-based approaches.

In this paper, we study the iteration-complexity of no-regret learning in games under the constraint that the equilibrium reached approximates the optimal social welfare (or some other natural objective); henceforth, we will simply refer to such equilibria as near-optimal. Taking a step back, there has been tremendous interest in understanding the performance of no-regret learning in terms of welfare, primarily stemming from the price of anarchy literature (e.g.[Rou15, Blu+08]), but here we ask an entirely different but equally fundamental question:

How many iterations are needed so that no-regret players converge to a near-optimal (approximate) equilibrium?

In terms of proving a lower bound, one natural approach revolves around the premise that players initially possess no information about the game, and in each iteration they receive only some limited utility feedback. We argue that there are certain caveats to such an approach. First, it does not apply to the usual centralized model of computation where the underlying game is given as part of the input. Furthermore, even in decentralized settings there is often a centralized party endeavoring to intervene and guide players to desirable outcomes [Mgu+19, KYV20, Li+20, Liu+22, BBM13, Bal+14]. Many models have been proposed that differ based on the amount of information gathered by the centralized party, as well as the way communication occurs between the different entities; each such model is arguably reasonable depending on the application. The question thus is how to come up with a lower bound that is not brittle to assumptions regarding the way information is distributed, and thereby applies to all such settings.

We address this by resorting to computational lower bounds, thereby ruling out fast convergence to a near-optimal equilibrium even when the entire game is known in advance—subsuming the so-called full feedback setting (recalled in Section 2)—and players can fully coordinate; at first glance, it might seem counterintuitive that non-trivial lower bounds can be established under such permissive assumptions. In particular, we focus on perhaps the simplest class of games for which such questions become meaningful: two-player (general-sum) games represented in normal form; since we are aiming to prove lower bounds, concentrating on a simple class of games only makes the result stronger.

1.1 Our results

We establish tight computational lower bounds for the number of iterations needed for no-regret players to reach a near-optimal equilibrium in two-player games. To do so, a key observation that drives our approach is that no-regret learning produces, essentially by definition, a CCE with a particular structure: one expressed as a mixture (that is, a convex combination) of product distributions. In particular, TT rounds of learning results in a mixture of TT product distributions. We call such a distribution TT-sparse (Definition 2.1). In this context, our main contribution is to prove hardness results for the problem of computing a near-optimal TT-sparse CCE, which—by virtue of the observation above—immediately circumscribes the number of iterations for no-regret learning as well, even in the centralized model of computation. Even though this is a fundamental problem, to our knowledge, we are the first to examine its computational complexity as a function of TT.

One special case of this problem is well-understood: a near-optimal 11-sparse CCE is nothing other than a near-optimal Nash equilibrium, treated in the seminal work of [GZ89] (and subsequently extended by [CS08] and [KM18]), and shown to be \NP\NP-complete. On the other end of the spectrum, assuming that each player has nn available actions, it is easy to see that any correlated distribution—and in particular any CCE—is nn-sparse. In light of the well-known fact that the optimal CCE can be computed in polynomial time via a linear program (Proposition 2.3), we see that there is a phase transition dictated by the sparsity parameter. In fact, our first main result shows that attaining non-trivial sparsity in polynomial time is impossible (subject to \NP\P\neq\NP). Below, for an n×nn\times n two-player game 𝒢{\mathcal{G}}, we denote by OptimalSparseCCE(𝒢,T,ϵ,ϵ^)\textsc{OptimalSparseCCE}({\mathcal{G}},T,\epsilon,\hat{\epsilon}) the problem of computing a TT-sparse CCE with equilibrium gap at most ϵ\epsilon (in an additive sense; see Definition 2.2) and welfare at least 𝖮𝖯𝖳ϵ^\mathsf{OPT}-\hat{\epsilon}, where 𝖮𝖯𝖳\mathsf{OPT} is the welfare attained by the optimal TT-sparse CCE. (Further background is given later in Section 2.)

Theorem 1.1.

OptimalSparseCCE(𝒢,n1ϵ,nc,nc)\textsc{OptimalSparseCCE}({\mathcal{G}},n^{1-\epsilon},n^{-c},n^{-c}) with respect to n×nn\times n games is \NP\NP-hard for any constant ϵ>0\epsilon>0 and some constant cc.

This means that roughly nn iterations are needed for (computationally bounded) no-regret learners to converge to a CCE with 𝗉𝗈𝗅𝗒(1/n)\mathsf{poly}(1/n) equilibrium and optimality gap; that is, the trivial upper bound of nn is essentially the best one can hope for. Further, a slightly stronger complexity assumption precludes even a sparsity of n/2(𝗅𝗈𝗀n)1γn/2^{(\mathsf{log}n)^{1-\gamma}} for some constant γ>0\gamma>0 (Corollary 3.3). It is worth noting that TnT\coloneqq n iterations also represent a natural information-theoretic threshold: in the full feedback setting, to which our lower bounds readily apply, there is a trivial exploration protocol that enables each player to fully determine its own payoff matrix (by simply iterating over all rows or columns)—trivializing the problem at least in (two-player) zero-sum games.

In addition, Theorem 1.1 establishes a complexity separation between OptimalSparseCCE and SparseCCE—the latter problem lifts the welfare constraint imposed by the former. Namely, since SparseCCE(𝒢,T,0)\textsc{SparseCCE}({\mathcal{G}},T,0) is in \PPAD\PPAD even for T=1T=1 [Pap94], OptimalSparseCCE is harder (subject to \coNP\NP\coNP\neq\NP [JPY88]) for any sparsity Tn1ϵT\leq n^{1-\epsilon}.

Moreover, we strengthen Theorem 1.1 in two key aspects. First, we show that it applies under a broad class of objectives, beyond (utilitarian) welfare, which additionally includes the egalitarian social welfare and each player’s (individual) utility (Corollaries 3.13 and 3.12). Second, \NP-hardness persists for any multiplicative approximation to the objective (Corollary 3.14). The key construction behind those results, Theorem 3.11, implies similar hardness results for two other natural problems pertaining to sparse CCE: deciding uniqueness (Theorem 3.9), and determining existence after excluding certain (joint) action profiles (Theorem 3.10); those two latter problems do not hinge on any underlying objective.

Technical approach

Compared to Nash equilibria, the crux in proving lower bounds for sparse CCE lies in introducing correlation between the players. Many natural reductions designed for Nash equilibria are of little use even for sparsity T=2T=2, which partly explains why the complexity of sparse CCE remains poorly understood. The key challenge is to identify a basic construction that handles near-optimal TT-sparse CCE even when T1T\gg 1.

In this context, to prove Theorem 1.1, we extend the reduction of [GZ89] who proved \NP\NP-hardness only when T=1T=1. In particular, they came up with a reduction from the decision version of the maximum clique problem (MaxClique) to OptimalSparseCCE(𝒢,1,0,0)\textsc{OptimalSparseCCE}({\mathcal{G}},1,0,0). We establish a natural generalization of their reduction (Algorithm 1); namely, we show that computing a 2T2T-approximation to MaxClique polynomially reduces to OptimalSparseCCE(𝒢,T,nc,nc)\textsc{OptimalSparseCCE}({\mathcal{G}},T,n^{-c},n^{-c}) (Theorem 3.1). That is, the sparsity of the underlying CCE translates to a degradation in the resulting approximation factor. We are then able to rely on the celebrated inapproximability of MaxClique [Zuc07] (Theorem 2.4) to arrive at Theorem 1.1. The overall reduction has various new technical aspects, discussed in Section 3.1. The refinements to Theorem 1.1 described earlier are established by suitably adjusting this basic reduction (Sections 3.2 and 3.3).

Low-precision regime

So far, we have focused on the regime where both the equilibrium and the optimality gap scale as 𝗉𝗈𝗅𝗒(1/n)\mathsf{poly}(1/n)—a common setting when it comes to equilibrium computation. No-regret learning is often employed in the so-called low-precision regime, which we identify with ϵ,ϵ^1/\polylogn\epsilon,\hat{\epsilon}\geq 1/\polylog n. In that setting, even Nash equilibria admit a quasipolynomial-time algorithm [LMM03], and so one cannot hope to prove—barring major complexity breakthroughs—\NP-hardness results. Instead, following an earlier work by [HK11], we rely on the so-called planted clique conjecture from average-case complexity (2.5; Section 2 provides a self-contained overview). We are then able to show the following quasipolynomial lower bounds.

Theorem 1.2.

Assuming that 2.5 holds, the following problems require nΩ(𝗅𝗈𝗀n)n^{\Omega(\mathsf{log}n)} time with respect to n×nn\times n games:

  • OptimalSparseCCE(𝒢,T,(𝗅𝗈𝗀n)c,(𝗅𝗈𝗀n)c)\textsc{OptimalSparseCCE}({\mathcal{G}},T,(\mathsf{log}n)^{-c},(\mathsf{log}n)^{-c}) for any T=\polylognT=\polylog n and some constant c=c(T)c=c(T),

  • OptimalSparseCCE(𝒢,T,c,c)\textsc{OptimalSparseCCE}({\mathcal{G}},T,c,c) for any T=O(1)T=O(1) and some constant c=c(T)c=c(T).

The first lower bound is shown by relying on our previous construction behind Theorem 1.1. The second one, which concerns the more permissive regime in which ϵ,ϵ^=Θ(1)\epsilon,\hat{\epsilon}=\Theta(1), adapts the reduction of [HK11] pertaining to optimal Nash equilibria. We provide the technical details in Section 4.

1.2 Further related work

The notion of a sparse CCE—a mixture of product distributions (Definition 2.1)—was recently studied by [FGK23] in the context of Markov (aka. stochastic) games to rule out the existence of polynomial-time no-regret algorithms—with respect to potentially non-Markovian deviations (see also the work of [PR24]). This stands in contrast to games represented in normal form, where the existence of efficient no-regret algorithms has been long known tracing back to [Bla56]. Yet, establishing non-trivial lower bounds for sparse CCE in normal-form games remains an open problem even for sparsity T=2T=2. It is worth highlighting that even though a CCE (without the sparsity constraint) can be computed exactly by solving a linear program [PR08], by far the most well-studied approach in the literature revolves around no-regret learning. This can be mostly attributed to the scalability, the minimal memory footprint, as well as the amenability to a distributed implementation of the latter approach, motivating the problem of sparse CCE. Besides this connection with no-regret learning, we argue that sparse CCE is a natural notion, worth examining in its own right, and ties to a long line of work on low-rank approximation in machine learning.

A related notion of sparsity imposes instead a bound on the number of nonzero elements of the distribution—that is, the size of its support. Unlike Definition 2.1, that latter notion is well-studied and understood (e.g., [BBP14]). It is clear that a distribution 𝝁\bm{\mu} with TT nonzero entries is TT-sparse per Definition 2.1, but the opposite does not hold in general. From the viewpoint of no-regret learning, proving lower bounds pertaining to distributions with small support translates to the setting where each player selects a pure strategy, while Definition 2.1 accounts for mixed strategies as well. To further elaborate on this difference, it is known that Ω(𝗅𝗈𝗀n/ϵ2)\Omega(\mathsf{log}n/\epsilon^{2}) nonzero elements in the support are information-theoretically necessary for even the existence of an ϵ\epsilon-Nash equilibrium in zero-sum games [FNS07], in turn implying that Ω(𝗅𝗈𝗀n/ϵ2)\Omega(\mathsf{log}n/\epsilon^{2}) iterations are needed when players select pure strategies. On the other hand, a consequence of the minimax theorem is that a 11-sparse equilibrium always exists, and can also be computed efficiently in such games via linear programming; this means that no superpolynomial computational lower bounds for no-regret learning in zero-sum games can be shown in mixed strategies.

A long-standing challenge in general-sum games is equilibrium selection: there could be a multiplicity of equilibria, and some are more reasonable than others. A common antidote—albeit certainly not the only one—is to identify an equilibrium maximizing the social welfare (or some other natural objective). This can be achieved in polynomial time even in multi-player (normal-form) games represented explicitly, which motivates investigating the complexity of OptimalSparseCCE—the focus of our work. However, this is no longer the case in succinct games, where maximizing welfare is typically \NP-hard [PR08] (cf. [BL15])—let alone sparsity constraints. This is also the case for two-player extensive-form games (e.g.[Zha+22]).

Another motivation for proving lower bounds revolving around TT-sparse CCE is that while many techniques that accelerate equilibrium computation rely on no-regret learning dynamics, they do not strictly comply with the traditional online nature of the framework. A notable example is alternation [Tam+15, WTP22, Cev+23], whereby players update their strategies sequentially—as opposed to simultaneous updates. Importantly, such techniques are captured through sparse CCE. Beyond computational considerations, it is worth pointing out an orthogonal line of work that has focused on query complexity aspects of (coarse) correlated equilibria (e.g.[GR16, BB15, Mai+23, GK23], and references therein).

2 Preliminaries

In this section, we introduce some necessary background on games and pertinent equilibrium concepts (Section 2.1), as well as the problem of approximating the maximum clique of a graph and the related planted clique problem (Section 2.2).

Notation

We use boldface letters, such as 𝒙\bm{x} and 𝒚\bm{y}, to denote vectors. Matrices are represented using boldface capital letters, such as 𝐑\mathbf{R} and 𝐂\mathbf{C}. In particular, 𝐈nn×n\mathbf{I}_{n}\in{\mathbb{R}}^{n\times n} denotes the identity matrix, while 𝟎nn×n\mathbf{0}_{n}\in{\mathbb{R}}^{n\times n} denotes the all-zeroes matrix. We use subscripts to access the coordinates of a vector or the entries of a matrix. Superscripts (together with parantheses) are typically reserved for the discrete time index. For a vector 𝒙\bm{x}, we use 𝒙\|\bm{x}\| to denote its (Euclidean) 2\ell_{2} norm and 𝒙1\|\bm{x}\|_{1} for the 1\ell_{1} norm. If 𝒙n+m\bm{x}\in{\mathbb{R}}^{n+m}, for n,m{1,2,}n,m\in{\mathbb{N}}\coloneqq\{1,2,\dots\}, we let n𝒙in(𝒙1,,𝒙n){\mathbb{R}}^{n}\ni\bm{x}_{i\leq n}\coloneqq(\bm{x}_{1},\dots,\bm{x}_{n}) and m𝒙in+1(𝒙n+1,,𝒙n+m){\mathbb{R}}^{m}\ni\bm{x}_{i\geq n+1}\coloneqq(\bm{x}_{n+1},\dots,\bm{x}_{n+m}). For TT\in{\mathbb{N}}, we use the shorthand notation [T]{1,,T}[T]\coloneqq\{1,\dots,T\}. For simplicity, we sometimes employ the notation O(),Ω(),Θ()O(\cdot),\Omega(\cdot),\Theta(\cdot) to suppresses (absolute) constants.

2.1 Two-player games and equilibrium concepts

We first provide the definition of a sparse coarse correlated equilibrium (CCE) (Definitions 2.1 and 2.2). Below and throughout, we focus on two-player games, but those definitions can be readily generalized to multi-player games as well (e.g.[CL06]).

Two-player games

By convention, we refer to the players as Player xx (for the “row” player) and Player yy (for the “column” player). We operate under the usual normal form representation of a (two-player) game. Here, Player xx and Player yy have a finite (and nonempty) set of actions 𝒜x{\mathcal{A}}_{x} and 𝒜y{\mathcal{A}}_{y}, respectively. The utility of Player xx and Player yy under a pair of actions (ax,ay)𝒜x×𝒜y(a_{x},a_{y})\in{\mathcal{A}}_{x}\times{\mathcal{A}}_{y} is given by 𝐑ax,ay\mathbf{R}_{a_{x},a_{y}} and 𝐂ax,ay\mathbf{C}_{a_{x},a_{y}}, respectively, where 𝐑,𝐂𝒜x×𝒜y\mathbf{R},\mathbf{C}\in{\mathbb{Q}}^{{\mathcal{A}}_{x}\times{\mathcal{A}}_{y}} are the payoff matrices of the game, which are given as part of the input. Players can randomize by specifying a mixed strategy—a probability distribution over their set of actions. For a pair of mixed strategies (𝒙,𝒚)Δ(𝒜x)×Δ(𝒜y)(\bm{x},\bm{y})\in\Delta({\mathcal{A}}_{x})\times\Delta({\mathcal{A}}_{y}), the (expected) utility of Player xx and Player yy is given by 𝔼(ax,ay)(𝒙,𝒚)𝐑ax,ay=𝒙,𝐑𝒚{\mathbb{E}}_{(a_{x},a_{y})\sim(\bm{x},\bm{y})}\mathbf{R}_{a_{x},a_{y}}=\langle\bm{x},\mathbf{R}\bm{y}\rangle and 𝔼(ax,ay)(𝒙,𝒚)𝐂ax,ay=𝒙,𝐂𝒚{\mathbb{E}}_{(a_{x},a_{y})\sim(\bm{x},\bm{y})}\mathbf{C}_{a_{x},a_{y}}=\langle\bm{x},\mathbf{C}\bm{y}\rangle, respectively.

Sparse CCE

We are now ready to introduce the notion of a sparse CCE. In the sequel, we will denote by 𝒙𝒚=𝒙𝒚\bm{x}\otimes\bm{y}=\bm{x}\bm{y}^{\top} the outer (tensor) product of 𝒙\bm{x} and 𝒚\bm{y}.

Definition 2.1 (Sparse (correlated) distributions).

Let 𝝁Δ(𝒜x×𝒜y)\bm{\mu}\in\Delta({\mathcal{A}}_{x}\times{\mathcal{A}}_{y}) be a (correlated) distribution supported on 𝒜x×𝒜y{\mathcal{A}}_{x}\times{\mathcal{A}}_{y}. We say that 𝝁\bm{\mu} is TT-sparse, for TT\in{\mathbb{N}}, if there exist 𝒙(1),,𝒙(T)Δ(𝒜x)\bm{x}^{(1)},\dots,\bm{x}^{(T)}\in\Delta({\mathcal{A}}_{x}), 𝒚(1),,𝒚(T)Δ(𝒜y)\bm{y}^{(1)},\dots,\bm{y}^{(T)}\in\Delta({\mathcal{A}}_{y}) and (α(1),,α(T))Δ([T])(\alpha^{(1)},\dots,\alpha^{(T)})\in\Delta([T]) such that 𝝁=t=1Tα(t)(𝒙(t)𝒚(t))\bm{\mu}=\sum_{t=1}^{T}\alpha^{(t)}(\bm{x}^{(t)}\otimes\bm{y}^{(t)}); that is, 𝝁\bm{\mu} is a mixture of TT product distributions. If α(1)=α(2)===α(T)=1/T\alpha^{(1)}=\alpha^{(2)}=\dots=\dots=\alpha^{(T)}=1/T, we say that 𝝁\bm{\mu} is uniform TT-sparse.

It is worth noting here a related notion of sparsity which imposes a bound on the support of 𝝁\bm{\mu}, discussed earlier in Section 1.2. We next recall the definition of a coarse correlated equilibrium.

Definition 2.2 (Coarse correlated equilibrium).

A (correlated) distribution 𝝁Δ(𝒜x×𝒜y)\bm{\mu}\in\Delta({\mathcal{A}}_{x}\times{\mathcal{A}}_{y}) is an ϵ\epsilon-coarse correlated equilibrium (ϵ\epsilon-CCE) if for any deviation (ax,ay)𝒜x×𝒜y(a_{x}^{\prime},a_{y}^{\prime})\in{\mathcal{A}}_{x}\times{\mathcal{A}}_{y},

𝔼(ax,ay)𝝁[𝐑ax,ay]\displaystyle{\mathbb{E}}_{(a_{x},a_{y})\sim\bm{\mu}}[\mathbf{R}_{a_{x},a_{y}}] 𝔼(,ay)𝝁[𝐑ax,ay]ϵ,\displaystyle\geq{\mathbb{E}}_{(\cdot,a_{y})\sim\bm{\mu}}[\mathbf{R}_{a_{x}^{\prime},a_{y}}]-\epsilon, (1)
𝔼(ax,ay)𝝁[𝐂ax,ay]\displaystyle{\mathbb{E}}_{(a_{x},a_{y})\sim\bm{\mu}}[\mathbf{C}_{a_{x},a_{y}}] 𝔼(ax,)𝝁[𝐂ax,ay]ϵ.\displaystyle\geq{\mathbb{E}}_{(a_{x},\cdot)\sim\bm{\mu}}[\mathbf{C}_{a_{x},a_{y}^{\prime}}]-\epsilon. (2)

A 0-CCE will be referred to as a CCE. Coarse correlated equilibria relax the notion of a correlated equilibrium [Aum74]. In the latter, Player xx can deviate with respect to any possible function ϕx:𝒜x𝒜x\phi_{x}:{\mathcal{A}}_{x}\to{\mathcal{A}}_{x} (and similarly for Player yy); that is, axa_{x}^{\prime} and aya_{y}^{\prime} in (1) and (2) are to be replaced with ϕx(ax)\phi_{x}(a_{x}) and ϕy(ay)\phi_{y}(a_{y}), respectively. We further remark that a 11-sparse CCE is, by definition, a Nash equilibrium (in particular, when 𝝁\bm{\mu} is supported on a single element, we have a pure Nash equilibrium).

Sparse equilibria and regret

There is an immediate but important connection between sparse CCE and the framework of no-regret learning. Namely, suppose that at every time t[T]t\in[T] players select strategies 𝒙(t)Δ(𝒜x)\bm{x}^{(t)}\in\Delta({\mathcal{A}}_{x}) and 𝒚(t)Δ(𝒜y)\bm{y}^{(t)}\in\Delta({\mathcal{A}}_{y}). In the full feedback setting, the players at time tt receive as feedback 𝒖x(t)𝐑𝒚(t)\bm{u}_{x}^{(t)}\coloneqq\mathbf{R}\bm{y}^{(t)} and 𝒖y(t)𝐂𝒙(t)\bm{u}_{y}^{(t)}\coloneqq\mathbf{C}^{\top}\bm{x}^{(t)}, respectively.111In online learning, it is customary to assume that the utility feedback is revealed to the players upon selecting their strategies, but our results do not rest on such assumptions (see Section 1.2 for a further discussion). The regret of Player xx is defined as 𝖱𝖾𝗀xTmax𝒙Δ(𝒜x)t=1T𝒙𝒙(t),𝒖x(t)\mathsf{Reg}^{T}_{x}\coloneqq\max_{\bm{x}^{\star}\in\Delta({\mathcal{A}}_{x})}\sum_{t=1}^{T}\langle\bm{x}^{\star}-\bm{x}^{(t)},\bm{u}_{x}^{(t)}\rangle, and similarly for Player yy. The key connection now, which follows readily from the definitions, is that no-regret learning produces a uniform TT-sparse max(𝖱𝖾𝗀xT/T,𝖱𝖾𝗀yT/T)\max(\mathsf{Reg}_{x}^{T}/T,\mathsf{Reg}_{y}^{T}/T)-CCE. As a concrete example, when both players employ multiplicative weights update, one obtains a uniform O(𝗅𝗈𝗀(|𝒜|)/ϵ2)O(\mathsf{log}(|{\mathcal{A}}|)/\epsilon^{2})-sparse ϵ\epsilon-CCE, where |𝒜|max(|𝒜x|,|𝒜y|)|{\mathcal{A}}|\coloneqq\max(|{\mathcal{A}}_{x}|,|{\mathcal{A}}_{y}|).

Further, the non-uniform version of Definition 2.1 corresponds to a weighted notion of regret in which a weight α(t)\alpha^{(t)} is attached to the ttth summand, with (α(1),,α(T))Δ([T])(\alpha^{(1)},\dots,\alpha^{(T)})\in\Delta([T]). The latter has received ample of interest in prior work (e.g.[Abe+18, BS19, ZLB22]), not least because it typically performs better in practice. Of course, precluding efficient computation of sparse CCE under non-uniform mixtures can only be stronger.

Social welfare

The expected (social) welfare, denoted by 𝖲𝖶()\operatorname{\mathsf{SW}}(\cdot), under a pair of mixed strategies (𝒙,𝒚)Δ(𝒜x)×Δ(𝒜y)(\bm{x},\bm{y})\in\Delta({\mathcal{A}}_{x})\times\Delta({\mathcal{A}}_{y}) is the sum of the players’ utilities: 𝖲𝖶(𝒙𝒚)𝒙,(𝐑+𝐂)𝒚\operatorname{\mathsf{SW}}(\bm{x}\otimes\bm{y})\coloneqq\langle\bm{x},(\mathbf{R}+\mathbf{C})\bm{y}\rangle. More generally, the welfare of a distribution 𝝁\bm{\mu} on 𝒜x×𝒜y{\mathcal{A}}_{x}\times{\mathcal{A}}_{y} is 𝖲𝖶(𝝁)𝔼(ax,ay)𝝁[𝐑ax,ay+𝐂ax,ay]\operatorname{\mathsf{SW}}(\bm{\mu})\coloneqq{\mathbb{E}}_{(a_{x},a_{y})\sim\bm{\mu}}[\mathbf{R}_{a_{x},a_{y}}+\mathbf{C}_{a_{x},a_{y}}]. While welfare is perhaps the most common objective considered in the literature, some of our lower bounds apply to other natural objectives as well (Corollaries 3.12 and 3.13).

Without any essential loss, we will henceforth switch our attention to games in which |𝒜x|=|𝒜y||{\mathcal{A}}_{x}|=|{\mathcal{A}}_{y}|; this can always be enforced by introducing “dummy” actions. For notational convenience, we will also let 𝒜x,𝒜y{1,2,,n}{\mathcal{A}}_{x},{\mathcal{A}}_{y}\coloneqq\{1,2,\dots,n\}; that is, the action of each player is represented via an integer (the corresponding player will be made clear from the context). Such a game 𝒢{\mathcal{G}} will oftentimes be referred to as n×nn\times n. We now highlight the following fact.

Proposition 2.3.

For any n×nn\times n (two-player) game, there is a polynomial-time algorithm that computes a CCE maximizing the welfare. Further, any distribution 𝛍Δ([n]×[n])\bm{\mu}\in\Delta([n]\times[n]) is nn-sparse.

Indeed, regarding the first part of the claim, it is well-known that computing a CCE maximizing the welfare can be expressed as a (polynomial) linear program—this holds more generally in normal-form games represented explicitly and under any linear objective. To justify the claim that any correlated distribution 𝝁\bm{\mu} is nn-sparse, we define, for all t[n]t\in[n], α(t)=j=1n𝝁t,j\alpha^{(t)}=\sum_{j=1}^{n}\bm{\mu}_{t,j}; 𝒙i(t)=1\bm{x}^{(t)}_{i}=1 for i=ti=t and 0 otherwise; and 𝒚(t)(𝝁t,1,,𝝁t,n)\bm{y}^{(t)}\propto(\bm{\mu}_{t,1},\dots,\bm{\mu}_{t,n}). By definition, t=1nα(t)(𝒙(t)𝒚(t))=𝝁\sum_{t=1}^{n}\alpha^{(t)}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})=\bm{\mu}, thereby implying that 𝝁\bm{\mu} is nn-sparse. On the other hand, 𝝁\bm{\mu} cannot necessarily be expressed as a uniform mixture of nn product distributions (e.g., when 𝝁\bm{\mu} is supported only on the diagonal).

2.2 Maximum clique and planted clique

Consider an undirected nn-node graph G=(V,E)G=(V,E), where we let V[n]V\coloneqq[n]. A clique in GG is a subset of the nodes KVK\subseteq V such that for {i,j}E\{i,j\}\in E for all i,jKi,j\in K with iji\neq j. The maximum clique problem (MaxClique) takes as input an nn-node graph, and asks for the maximum clique in GG. The decision version of MaxClique was featured in the original list of \NP\NP-complete problems compiled by [Kar72]. Accordingly, commencing with a celebrated connection with probabilistic checkable proofs (PCPs), its (in)approximability received tremendous attention in theoretical computer science (e.g.[Aro+98, BS94, BGS98, Fei+96, Hås99, Zuc07]), culminating in the hardness results of [Hås99] and [Zuc07] stated below. We recall that if ω(G)\omega(G) is the size of the maximum clique in GG, a (multiplicative) approximation of ρ1\rho\geq 1 is a clique of size at least ω(G)/ρ\omega(G)/\rho. For our purposes, we will make use of the following hardness result [Zuc07, Theorem 1.1].

Theorem 2.4 ([Zuc07], \citeyearZuckerman07:Linear).

It is \NP-hard to approximate MaxClique with respect to an nn-node graph to a factor of n1ϵn^{1-\epsilon} for any constant ϵ>0\epsilon>0.

The trivial approximation factor for MaxClique is nn (since ω(G)n\omega(G)\leq n), and so Theorem 2.4 tells us that there is no algorithm significantly better than the trivial one. [Zuc07] also showed, by derandomizing the earlier lower bound of [Kho01], that it is 𝖭𝖯~\mathsf{N}\widetilde{\mathsf{P}}-hard to approximate MaxClique to a factor of n/2(𝗅𝗈𝗀n)1γn/2^{(\mathsf{log}n)^{1-\gamma}} for some constant γ>0\gamma>0 [Zuc07, Theorem 1.3]. 𝖭𝖯~\mathsf{N}\widetilde{\mathsf{P}} here is the quasipolynomial analogue of \NP\NP, where quasipolynomial stands for 2\polylogn2^{\polylog n}.

Planted clique problem

Related to MaxClique, the so-called planted (aka. hidden) clique problem (PlantedClique) was first studied independently by [Jer92] and [Kuč95]. Here, a random graph G=G(n,12,k)G=G(n,\frac{1}{2},k) is constructed as follows. First, an arbitrary subset of k=k(n)[n]k=k(n)\in[n] nodes is selected and a clique is placed on them. Then, every other edge is included independently with probability 12\frac{1}{2}, and the goal is to recover the planted clique. In a random Erdős-Rényi graph the size of the largest clique is (2+on(1))𝗅𝗈𝗀n(2+o_{n}(1))\mathsf{log}n with high probability (w.h.p.),222We say that an event \mathcal{E} occurs with high probability if []11/nΘ(1)\operatorname{\mathbb{P}}[\mathcal{E}]\geq 1-1/n^{\Theta(1)}. and so when k=(2+ϵ)𝗅𝗈𝗀nk=(2+\epsilon)\mathsf{log}n, finding the planted clique is information-theoretically possible by identifying the maximum clique of GG. Now, for k(2+ϵ)𝗅𝗈𝗀nk\geq(2+\epsilon)\mathsf{log}n, it is not hard to see that the planted clique can be identified in quasipolynomial time nO(𝗅𝗈𝗀n)n^{O(\mathsf{log}n)}; the planted clique conjecture postulates that there is no significantly faster algorithm. In particular, we will use the following form of the conjecture.

Conjecture 2.5.

Solving PlantedClique w.h.p. when k=\polylognk=\polylog n requires nΩ(𝗅𝗈𝗀n)n^{\Omega(\mathsf{log}n)} time.

In the precise sense of Lemma 4.1, PlantedClique becomes easier as kk grows; the current threshold for polynomial-time algorithms is k=Θ(n)k=\Theta(\sqrt{n}) [AKS98]. It is believed—and there is strong evidence for it [Bar+19]—that no polynomial-time algorithm exists when knk\ll\sqrt{n}; the weaker version of 2.5 will suffice for our purposes. A number of prior works from diverse areas have relied on the planted clique conjecture to prove different hardness results, such as testing kk-wise independence [Alo+07], finding clusters in social networks [Bal+13], public-key encryption [ABW10], and sparse PCA [BR13]. Closely related to our work is the result of [HK11] pertaining to the complexity of computing Nash equilibria with near-optimal welfare (see also [MV09, ABC13]), which we extend to OptimalSparseCCE (Theorem 1.2).

3 The complexity of computing near-optimal sparse CCE

In this section, we establish our main results concerning the complexity of OptimalSparseCCE and other related problems in the 𝗉𝗈𝗅𝗒(1/n)\mathsf{poly}(1/n) regime. The low-precision regime is treated in Section 4.

3.1 The basic reduction from approximate MaxClique

We first establish a reduction from approximating MaxClique to the problem of computing sparse (approximate) CCE with near-optimal welfare (Algorithm 1). In what follows, we recall that we refer to the latter computational problem as OptimalSparseCCE(𝒢,T,ϵ,ϵ^)\textsc{OptimalSparseCCE}({\mathcal{G}},T,\epsilon,\hat{\epsilon}), which takes as input a (two-player) game 𝒢{\mathcal{G}} in normal form; a sparsity parameter TT\in{\mathbb{N}} per Definition 2.1; the (CCE) equilibrium gap ϵ\epsilon per Definition 2.2; and the optimality gap ϵ^\hat{\epsilon}, also in an additive sense (multiplicative (in)approximability is the subject of Corollary 3.14). Then, it returns (deterministically) a TT-sparse ϵ\epsilon-CCE with welfare at least 𝖮𝖯𝖳ϵ^\mathsf{OPT}-\hat{\epsilon}, where 𝖮𝖯𝖳\mathsf{OPT} is the welfare of the optimal TT-sparse CCE. Before proceeding, it is worth noting that without the equilibrium constraint the problem becomes trivial: one can just return the maximum entry of 𝐑+𝐂\mathbf{R}+\mathbf{C}, which is of course 11-sparse.

The reduction

Suppose that we are given as input an undirected graph G=([n],E)G=([n],E) with adjacency matrix 𝐀=𝐀(G)n×n\mathbf{A}=\mathbf{A}(G)\in{\mathbb{R}}^{n\times n}, defined such that

𝐀i,j={1if {i,j}E(i=j), and0otherwise.\mathbf{A}_{i,j}=\begin{cases}1&\text{if }\{i,j\}\in E\lor(i=j),\text{ and}\\ 0&\text{otherwise}.\end{cases}

(It is convenient to assume that each entry in the diagonal of 𝐀\mathbf{A} is also 11.) The following reduction mostly relies on the earlier construction of [GZ89], but with some new, distinct ingredients. In particular, our reduction is given as Algorithm 1. Algorithm 1 takes as input GG, and constructs for each kk in Algorithm 1 a two-player game (Algorithm 1) defined by the (2n)×(2n)(2n)\times(2n) payoff matrices

(2n)×(2n)𝐑12(𝐀+γ𝐈nk𝐈nk𝐈n𝟎n×n) and (2n)×(2n)𝐂12(𝐀+γ𝐈nk𝐈nk𝐈n𝟎n×n),{\mathbb{R}}^{(2n)\times(2n)}\ni\mathbf{R}\coloneqq\frac{1}{2}\begin{pmatrix}\mathbf{A}+\gamma\mathbf{I}_{n}&-k\mathbf{I}_{n}\\ k\mathbf{I}_{n}&\mathbf{0}_{n\times n}\end{pmatrix}\text{ and }{\mathbb{R}}^{(2n)\times(2n)}\ni\mathbf{C}\coloneqq\frac{1}{2}\begin{pmatrix}\mathbf{A}+\gamma\mathbf{I}_{n}&k\mathbf{I}_{n}\\ -k\mathbf{I}_{n}&\mathbf{0}_{n\times n}\end{pmatrix}, (3)

where we recall that 𝐈nn×n\mathbf{I}_{n}\in{\mathbb{R}}^{n\times n} denotes the identity matrix, 𝟎n×nn×n\mathbf{0}_{n\times n}\in{\mathbb{R}}^{n\times n} denotes the all-zeroes matrix, and parameter γ1\gamma\ll 1 is specified in Algorithm 1.333One can normalize the above payoff matrices so that all entries are in [1,1][-1,1], which is typically assumed when additive approximations are considered. All our hardness results apply as stated under that assumption. (An example of 𝒢=𝒢(G){\mathcal{G}}={\mathcal{G}}(G) for the 44-node graph GG of Figure 2 is given in Figure 2.) Then, Algorithm 1 invokes OptimalSparseCCE(𝒢,T,ϵ,ϵ^)\textsc{OptimalSparseCCE}({\mathcal{G}},T,\epsilon,\hat{\epsilon}) for a suitable choice of ϵ\epsilon and ϵ^\hat{\epsilon} (Algorithm 1). It is worth pointing out that game (3) is symmetric, in that 𝐑=𝐂\mathbf{R}=\mathbf{C}^{\top}, which will only make the lower bound stronger.

1234
Figure 1: A 44-node graph GG.
1 2 3 4
1 1+γ1+\gamma 1 1 1 k-k 0 0 0
2 1 1+γ1+\gamma 1 0 0 k-k 0 0
3 1 1 1+γ1+\gamma 1 0 0 k-k 0
4 1 0 1 1+γ1+\gamma 0 0 0 k-k
kk 0 0 0 0 0 0 0
0 kk 0 0 0 0 0 0
0 0 kk 0 0 0 0 0
0 0 0 kk 0 0 0 0
Figure 2: The payoff matrix 2𝐑=2𝐑(G)2\cdot\mathbf{R}=2\cdot\mathbf{R}(G) per (3).
Data: nn-node graph GG, sparsity TT\in{\mathbb{N}}
Result: A 2T2T-approximation to the maximum clique of GG
1 Set K{i}K\leftarrow\{i\} for any i[n]i\in[n]
2 for k=2T,3T,,n/TTk=2T,3T,\dots,\lfloor n/T\rfloor T  do
3       Set γ<1402k6(k+1)2\gamma<\frac{1}{40^{2}k^{6}(k+1)^{2}}
4       Construct the (2n)×(2n)(2n)\times(2n) (two-player) game 𝒢=𝒢(k,γ){\mathcal{G}}={\mathcal{G}}(k,\gamma) per (3)
5       Set the optimality gap ϵ^γ2T/2\hat{\epsilon}\leftarrow\gamma^{2}T/2 and the equilibrium gap ϵkγ/2\epsilon\leftarrow k\gamma/2
6       t=1Tα(t)(𝒙(t)𝒚(t))OptimalSparseCCE(𝒢,T,ϵ,ϵ^)\sum_{t=1}^{T}\alpha^{(t)}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})\leftarrow\textsc{OptimalSparseCCE}({\mathcal{G}},T,\epsilon,\hat{\epsilon})
7       Set 𝒙^in(t)=𝒙in(t)/i=1n𝒙i(t)\widehat{\bm{x}}_{i\leq n}^{(t)}=\nicefrac{{\bm{x}^{(t)}_{i\leq n}}}{{\sum_{i=1}^{n}\bm{x}_{i}^{(t)}}} and 𝒚^jn(t)=𝒚jn(t)/j=1n𝒚j(t)\widehat{\bm{y}}_{j\leq n}^{(t)}=\nicefrac{{\bm{y}^{(t)}_{j\leq n}}}{{\sum_{j=1}^{n}\bm{y}_{j}^{(t)}}} for all t[T]t\in[T]
8       Determine targmax1tT(α(t))t^{\star}\in\arg\max_{1\leq t\leq T}(\alpha^{(t)})
9       Let S{i[n]:𝒙^i(t)140kγ}S\leftarrow\{i\in[n]:\widehat{\bm{x}}_{i}^{(t^{\star})}\geq\frac{1}{\ell}-40\ell k\sqrt{\gamma}\}, where α(t)k\ell\coloneqq\lfloor\alpha^{(t^{\star})}k\rfloor
10       if  SS induces an \ell-clique in GG then
11             Set KSK\leftarrow S
12       end if
13      
14 end for
Output: K[n]K\subseteq[n]
Algorithm 1 Reducing (approximate) MaxClique to OptimalSparseCCE

It is clear that, with the possible exception of Algorithm 1, all other steps of Algorithm 1 can be immediately implemented in time 𝗉𝗈𝗅𝗒(n)\mathsf{poly}(n) (TT of course can be safely assumed to be polynomial in nn). The main claim regarding Algorithm 1 is summarized below.

Theorem 3.1.

For any nn-node graph GG and sparsity parameter TT\in{\mathbb{N}}, Algorithm 1 returns a 2T2T-approximation to the maximum clique of GG.

The proof of this theorem is the subject of Section 3.1.1 below. Assuming Theorem 3.1, we arrive at the following polynomial-time reduction from approximate MaxClique to OptimalSparseCCE.

Corollary 3.2.

There exists a polynomial-time reduction from 2T2T-approximate MaxClique to OptimalSparseCCE(𝒢,T,nc,nc)\textsc{OptimalSparseCCE}({\mathcal{G}},T,n^{-c},n^{-c}) for some constant cc.

Thus, combining with the inapproximability of MaxClique (Theorem 2.4), we obtain the following main implications, the first of which was stated earlier in the introduction.

See 1.1

Corollary 3.3.

OptimalSparseCCE(𝒢,n/2(𝗅𝗈𝗀n)1γ,nc,nc)\textsc{OptimalSparseCCE}({\mathcal{G}},n/2^{(\mathsf{log}n)^{1-\gamma}},n^{-c},n^{-c}) with respect to n×nn\times n games is 𝖭𝖯~\mathsf{N}\widetilde{\mathsf{P}}-hard for some constants γ>0\gamma>0 and cc.

To put those results into context, and as we pointed out earlier in Proposition 2.3, there is a polynomial-time algorithm for solving OptimalSparseCCE(𝒢,n,0,0)\textsc{OptimalSparseCCE}({\mathcal{G}},n,0,0) via linear programming.

3.1.1 Proof of Theorem 3.1

We split the main argument of the proof of Theorem 3.1 into Lemmas 3.4, 3.5, 3.6 and 3.7 below. Let us first provide a basic high-level overview of the approach for the special case of T=1T=1, which goes back to [GZ89]. Perhaps the first attempt that comes to mind when trying to reduce from MaxClique consists of using the adjacency matrix as the payoff matrix of each player. Of course, this reduction gives rise to uninteresting (pure) equilibria with optimal welfare: for an edge in GG, say {1,2}E\{1,2\}\in E, Player xx can select 11 and Player 22 can select 22. This is where the component k𝐈nk\mathbf{I}_{n} in (3) comes into play: it forces each player to select each action with at most a probability of 1/k\approx 1/k—for otherwise the other player could profitably deviate to that component. Further, since only the joint strategies corresponding to 𝐀\mathbf{A} have positive welfare, any Nash equilibrium with high welfare has to allocate most of its mass in that component. Anti-concentration, however, is still not enough, for the players may still end up supporting their strategies on a complete bipartite subgraph in GG (that is, each player on a different part). Incorporating the diagonal term γ𝐈n\gamma\mathbf{I}_{n} in the adjacency matrix turns out to address that issue as well, as it incentives players to select strategies that are close to each other. With that basic skeleton in mind, we next analyze the reduction under OptimalSparseCCE.

In what follows, we shall consider a fixed value of kk, and show that if GG contains a clique of size kk, then Algorithm 1 indeed outputs a clique of size k/Tk/T. We begin with a characterization of the set of optimal CCE of game (3), (𝐑,𝐂)(\mathbf{R},\mathbf{C}), as a function of the sparsity parameter TT.

Lemma 3.4.

Suppose that graph GG contains a clique of size kk\in{\mathbb{N}}. Then, assuming that k/Tk/T\in{\mathbb{N}}, there is a (uniform) TT-sparse CCE in (𝐑,𝐂)(\mathbf{R},\mathbf{C}) with welfare 1+γT/k1+\gamma T/k.

As a result, the welfare bound attained by TT-sparse CCE according to Lemma 3.4 improves as TT increases, which will be crucial for the upcoming argument. Indeed, focusing on the n×nn\times n submatrix of (3) corresponding to the adjacency matrix 𝐀\mathbf{A}, we observe that the diagonal entries have strictly larger utility than all other entries. The basic idea now is that one can use the multiple mixtures of the CCE in order to assign more probability mass in the diagonal elements than what is possible under a product distribution, without violating the equilibrium constraint. Namely, we partition the kk-clique into TT (disjoint) subsets, and every mixture is set to be the uniform distribution over the corresponding subset. In turn, this is reflected in attaining higher welfare. An instance of this reasoning when n=4=kn=4=k and T=2T=2 is illustrated in Figure 2 (although that graph does not admit a 44-clique). We now formalize this argument.

Proof of Lemma 3.4.

Let KVK\subseteq V be a clique in GG of size kk, which is assumed to exist. We construct a mixture of product distributions as follows. We partition KK into TT (disjoint) subsets K(1),K(2),,K(T)K^{(1)},K^{(2)},\dots,K^{(T)}, each of size k/Tk/T. For every t[T]t\in[T], we define 𝒙(t)\bm{x}^{(t)} and 𝒚(t)\bm{y}^{(t)} to be the uniform distributions over K(t)K^{(t)}. Then, we let 𝝁1Tt=1T𝒙(t)𝒚(t)\bm{\mu}\coloneqq\frac{1}{T}\sum_{t=1}^{T}\bm{x}^{(t)}\otimes\bm{y}^{(t)}.

We first claim that 𝝁\bm{\mu} attains a welfare of 1+γT/k1+\gamma T/k. Indeed, by the above definition of (𝒙(t),𝒚(t))(\bm{x}^{(t)},\bm{y}^{(t)}), (i,j)(𝒙(t),𝒚(t))[i=j]=T/k\operatorname{\mathbb{P}}_{(i,j)\sim(\bm{x}^{(t)},\bm{y}^{(t)})}[i=j]=T/k for any t[T]t\in[T]. So, 𝖲𝖶(𝒙(t)𝒚(t))=1+γT/k\operatorname{\mathsf{SW}}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})=1+\gamma T/k, where we used the fact that K(t)[n]K^{(t)}\subseteq[n] forms a clique on GG (by construction). In turn, this implies that 𝖲𝖶(𝝁)=1Tt=1T𝖲𝖶(𝒙(t)𝒚(t))=1+γT/k\operatorname{\mathsf{SW}}(\bm{\mu})=\frac{1}{T}\sum_{t=1}^{T}\operatorname{\mathsf{SW}}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})=1+\gamma T/k.

Next, we show that 𝝁\bm{\mu} is a CCE. By symmetry, it suffices to consider deviations of Player yy. We first observe that the marginal of 𝝁\bm{\mu} with respect to Player xx, 1Tt=1T𝒙(t)\frac{1}{T}\sum_{t=1}^{T}\bm{x}^{(t)}, is the uniform distribution over the clique KK. Thus, any deviation of Player yy to an action j[n]j\in[n] can yield an (expected) utility of at most 12(1+γ/k)\frac{1}{2}(1+\gamma/k); this is at most the utility obtained under 𝝁\bm{\mu}, which we saw was 12(1+γT/k)\frac{1}{2}(1+\gamma T/k) (since the players receive the same utility when the game is restricted on 𝐀\mathbf{A}). Further, any deviation to an action n+1j2nn+1\leq j\leq 2n can lead to a utility of at most 12\frac{1}{2}. This is again strictly worse than what obtained under 𝝁\bm{\mu} since γ>0\gamma>0, completing the proof. ∎

It is interesting to note that, when T2T\geq 2, the distribution described above is not a correlated equilibrium (but merely a coarse correlated equilibrium). Let us consider, for example, the extreme scenario where T=kT=k, in which case 𝝁\bm{\mu} is supported solely on the diagonal elements of the clique KK. Considering Player yy, we now see that any deviation jn+jj\mapsto n+j for jKj\in K secures a utility of k2\frac{k}{2}, which is much higher than the utility obtained under 𝝁\bm{\mu}. The difference here is that Player yy receives its own component of 𝝁\bm{\mu} before contemplating a deviation; but since 𝝁\bm{\mu} is supported only on the diagonal, Player yy knows the action of Player xx with probability 11.

Next, to extract interesting information about GG from a CCE in (3), we need to ensure that players are typically playing on the n×nn\times n submatrix corresponding to 𝐀\mathbf{A}. To do so, the observation is that upon excluding that submatrix, the rest of the game is zero-sum, and hence the attained welfare is nullified. But we saw earlier in Lemma 3.4 that there is a CCE with high welfare, and so a near-optimal CCE cannot afford to assign too much mass outside of that n×nn\times n component, as reflected in the lemma below.

Lemma 3.5.

Consider a TT-sparse ϵ\epsilon-CCE 𝛍=t=1Tα(t)(𝐱(t)𝐲(t))\bm{\mu}=\sum_{t=1}^{T}\alpha^{(t)}(\bm{x}^{(t)}\otimes\bm{y}^{(t)}) of (𝐑,𝐂)(\mathbf{R},\mathbf{C}) with 𝖲𝖶(𝛍)1\operatorname{\mathsf{SW}}(\bm{\mu})\geq 1. For each t[T]t\in[T], we define 𝛍^=t=1Tα(t)(𝐱^(t)𝐲^(t))\widehat{\bm{\mu}}=\sum_{t=1}^{T}\alpha^{(t)}(\widehat{\bm{x}}^{(t)}\otimes\widehat{\bm{y}}^{(t)}), where 𝐱^in(t)𝐱in(t)/i=1n𝐱i(t)\widehat{\bm{x}}_{i\leq n}^{(t)}\coloneqq\nicefrac{{\bm{x}^{(t)}_{i\leq n}}}{{\sum_{i=1}^{n}\bm{x}_{i}^{(t)}}} and 𝐲^jn(t)𝐲jn(t)/j=1n𝐲j(t)\widehat{\bm{y}}_{j\leq n}^{(t)}\coloneqq\nicefrac{{\bm{y}^{(t)}_{j\leq n}}}{{\sum_{j=1}^{n}\bm{y}_{j}^{(t)}}}. Then, it holds that 𝛍^\widehat{\bm{\mu}} is an (ϵ+2kγ)(\epsilon+2k\gamma)-CCE of (𝐑,𝐂)(\mathbf{R},\mathbf{C}) with 𝖲𝖶(𝛍^)𝖲𝖶(𝛍)\operatorname{\mathsf{SW}}(\widehat{\bm{\mu}})\geq\operatorname{\mathsf{SW}}(\bm{\mu}). Further, by construction, 𝛍^\widehat{\bm{\mu}} is supported only on [n]×[n][n]\times[n].

We clarify that a distribution 𝝁Δ([2n]×[2n])\bm{\mu}\in\Delta([2n]\times[2n]) is said to be supported only on [n]×[n][n]\times[n] if 𝝁i,j=0\bm{\mu}_{i,j}=0 when i[n]i\notin[n] or j[n]j\notin[n]. We also note that the understanding above is that 𝒙^,𝒚^Δ2n\widehat{\bm{x}},\widehat{\bm{y}}\in\Delta^{2n}, although they are supported only on [n][n]. Further, if i=1n𝒙i(t)=0\sum_{i=1}^{n}\bm{x}_{i}^{(t)}=0, 𝒙^in(t)\widehat{\bm{x}}_{i\leq n}^{(t)} can be taken to be any point in Δn\Delta^{n}; the same convention is followed with respect to 𝒚^jn(t)\widehat{\bm{y}}_{j\leq n}^{(t)}.

Proof of Lemma 3.5.

We define δ(t)1i=1n𝒙i(t)j=1n𝒚j(t)\delta^{(t)}\coloneqq 1-\sum_{i=1}^{n}\bm{x}_{i}^{(t)}\sum_{j=1}^{n}\bm{y}_{j}^{(t)} for each t[T]t\in[T]. We will first show that t=1Tα(t)δ(t)γ\sum_{t=1}^{T}\alpha^{(t)}\delta^{(t)}\leq\gamma. Indeed, we can bound the welfare of each individual product distribution as

𝖲𝖶(𝒙(t)𝒚(t))=𝒙(t),(𝐑+𝐂)𝒚(t)=i=1nj=1n𝒙i(t)(𝐀i,j+γ𝟙{i=j})𝒚j(t)(1+γ)(1δ(t)),\operatorname{\mathsf{SW}}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})=\langle\bm{x}^{(t)},(\mathbf{R}+\mathbf{C})\bm{y}^{(t)}\rangle=\sum_{i=1}^{n}\sum_{j=1}^{n}\bm{x}_{i}^{(t)}(\mathbf{A}_{i,j}+\gamma\mathbbm{1}\{i=j\})\bm{y}_{j}^{(t)}\leq(1+\gamma)(1-\delta^{(t)}),

by definition of 𝐑\mathbf{R} and 𝐂\mathbf{C} in (3). Thus, we have

1𝖲𝖶(𝝁)=t=1Tα(t)𝖲𝖶(𝒙(t)𝒚(t))(1+γ)(1t=1Tα(t)δ(t)).1\leq\operatorname{\mathsf{SW}}(\bm{\mu})=\sum_{t=1}^{T}\alpha^{(t)}\operatorname{\mathsf{SW}}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})\leq(1+\gamma)\left(1-\sum_{t=1}^{T}\alpha^{(t)}\delta^{(t)}\right).

So, t=1Tα(t)δ(t)γ\sum_{t=1}^{T}\alpha^{(t)}\delta^{(t)}\leq\gamma. We now define 𝒙^Δ2n\widehat{\bm{x}}\in\Delta^{2n} such that 𝒙^in(t)=𝒙in(t)/i=1n𝒙i(t)\widehat{\bm{x}}_{i\leq n}^{(t)}=\bm{x}^{(t)}_{i\leq n}/\sum_{i=1}^{n}\bm{x}_{i}^{(t)} and 𝒚^Δ2n\widehat{\bm{y}}\in\Delta^{2n} such that 𝒚^jn(t)=𝒚jn(t)/j=1n𝒚j(t)\widehat{\bm{y}}_{j\leq n}^{(t)}=\bm{y}^{(t)}_{j\leq n}/\sum_{j=1}^{n}\bm{y}_{j}^{(t)} for all t[T]t\in[T]. Accordingly, we define 𝝁^t=1Tα(t)(𝒙^(t)𝒚^(t))\widehat{\bm{\mu}}\coloneqq\sum_{t=1}^{T}\alpha^{(t)}(\widehat{\bm{x}}^{(t)}\otimes\widehat{\bm{y}}^{(t)}); by construction, each product of 𝝁^\widehat{\bm{\mu}} is supported solely on [n]×[n][n]\times[n]. It is also clear that 𝖲𝖶(𝝁^)𝖲𝖶(𝝁)\operatorname{\mathsf{SW}}(\widehat{\bm{\mu}})\geq\operatorname{\mathsf{SW}}(\bm{\mu}) since, also by construction, 𝒙^i(t)𝒙i(t)\widehat{\bm{x}}_{i}^{(t)}\geq\bm{x}_{i}^{(t)} and 𝒚^j(t)𝒚j(t)\widehat{\bm{y}}_{j}^{(t)}\geq\bm{y}_{j}^{(t)} for all t[T]t\in[T] and i,j[n]i,j\in[n]. As a result, it suffices to bound the CCE gap of 𝝁^\widehat{\bm{\mu}}. To that end, we will consider only deviations by Player xx, as deviations by Player yy can be treated similarly. Since 𝝁\bm{\mu} is assumed to be an ϵ\epsilon-CCE, we have that for any 𝒙Δ2n\bm{x}\in\Delta^{2n},

t=1Tα(t)𝒙𝒙(t),𝐑𝒚(t)ϵ.\sum_{t=1}^{T}\alpha^{(t)}\langle\bm{x}-\bm{x}^{(t)},\mathbf{R}\bm{y}^{(t)}\rangle\leq\epsilon. (4)

Now, we bound

𝒙(t),𝐑𝒚(t)\displaystyle\langle\bm{x}^{(t)},\mathbf{R}\bm{y}^{(t)}\rangle i=1nj=1n𝒙i(t)𝐑i,j𝒚j(t)+i=n+12nj=1n𝒙i(t)𝐑i,j𝒚j(t)\displaystyle\leq\sum_{i=1}^{n}\sum_{j=1}^{n}\bm{x}_{i}^{(t)}\mathbf{R}_{i,j}\bm{y}_{j}^{(t)}+\sum_{i=n+1}^{2n}\sum_{j=1}^{n}\bm{x}_{i}^{(t)}\mathbf{R}_{i,j}\bm{y}_{j}^{(t)}
=12i=1nj=1n𝒙i(t)(𝐀i,j+γ𝟙{i=j})𝒚j(t)+k2(1i=1n𝒙i(t))(j=1n𝒚j(t))\displaystyle=\frac{1}{2}\sum_{i=1}^{n}\sum_{j=1}^{n}\bm{x}_{i}^{(t)}(\mathbf{A}_{i,j}+\gamma\mathbbm{1}\{i=j\})\bm{y}_{j}^{(t)}+\frac{k}{2}\left(1-\sum_{i=1}^{n}\bm{x}_{i}^{(t)}\right)\left(\sum_{j=1}^{n}\bm{y}_{j}^{(t)}\right)
12i=1nj=1n𝒙^i(t)(𝐀i,j+γ𝟙{i=j})𝒚^j(t)+k2(1i=1nj=1n𝒙i(t)𝒚j(t))\displaystyle\leq\frac{1}{2}\sum_{i=1}^{n}\sum_{j=1}^{n}\widehat{\bm{x}}_{i}^{(t)}(\mathbf{A}_{i,j}+\gamma\mathbbm{1}\{i=j\})\widehat{\bm{y}}_{j}^{(t)}+\frac{k}{2}\left(1-\sum_{i=1}^{n}\sum_{j=1}^{n}\bm{x}_{i}^{(t)}\bm{y}_{j}^{(t)}\right)
=𝒙^(t),𝐑𝒚^(t)+k2δ(t).\displaystyle=\langle\widehat{\bm{x}}^{(t)},\mathbf{R}\widehat{\bm{y}}^{(t)}\rangle+\frac{k}{2}\delta^{(t)}. (5)

Also, for any 𝒙Δ2n\bm{x}\in\Delta^{2n},

𝒙,𝐑𝒚(t)=𝒙,𝐑𝒚^(t)i=12nj=12n𝒙i𝐑i,j(𝒚j(t)𝒚^j(t))𝒙,𝐑𝒚^(t)kδ(t),\langle\bm{x},\mathbf{R}\bm{y}^{(t)}\rangle=\langle\bm{x},\mathbf{R}\widehat{\bm{y}}^{(t)}\rangle-\sum_{i=1}^{2n}\sum_{j=1}^{2n}\bm{x}_{i}\mathbf{R}_{i,j}(\bm{y}^{(t)}_{j}-\widehat{\bm{y}}^{(t)}_{j})\geq\langle\bm{x},\mathbf{R}\widehat{\bm{y}}^{(t)}\rangle-k\delta^{(t)}, (6)

where the inequality above uses the fact that 𝒚(t)𝒚^(t)1=2(1j=1n𝒚j(t))\|\bm{y}^{(t)}-\widehat{\bm{y}}^{(t)}\|_{1}=2(1-\sum_{j=1}^{n}\bm{y}_{j}^{(t)}). Indeed, when j=1n𝒚j(t)=0\sum_{j=1}^{n}\bm{y}_{j}^{(t)}=0, we have 𝒚(t)𝒚^(t)1=𝒚(t)1+𝒚^(t)1=2=2(1j=1n𝒚j(t))\|\bm{y}^{(t)}-\widehat{\bm{y}}^{(t)}\|_{1}=\|\bm{y}^{(t)}\|_{1}+\|\widehat{\bm{y}}^{(t)}\|_{1}=2=2(1-\sum_{j=1}^{n}\bm{y}_{j}^{(t)}); in the contrary case, we have

𝒚(t)𝒚^(t)1\displaystyle\|\bm{y}^{(t)}-\widehat{\bm{y}}^{(t)}\|_{1} =𝒚jn+1(t)1+𝒚jn(t)𝒚jn(t)j=1n𝒚j(t)1\displaystyle=\|\bm{y}^{(t)}_{j\geq n+1}\|_{1}+\left\|\bm{y}^{(t)}_{j\leq n}-\frac{\bm{y}^{(t)}_{j\leq n}}{\sum_{j=1}^{n}\bm{y}_{j}^{(t)}}\right\|_{1}
=1j=1n𝒚j(t)+(11j=1n𝒚j(t))j=1n𝒚j(t)=2(1j=1n𝒚j(t)).\displaystyle=1-\sum_{j=1}^{n}\bm{y}_{j}^{(t)}+\left(1-\frac{1}{\sum_{j=1}^{n}\bm{y}_{j}^{(t)}}\right)\sum_{j=1}^{n}\bm{y}_{j}^{(t)}=2\left(1-\sum_{j=1}^{n}\bm{y}_{j}^{(t)}\right).

Combining (5) and (6) with (4), it follows that for any 𝒙Δ2n\bm{x}\in\Delta^{2n},

t=1Tα(t)𝒙𝒙^(t),𝐑𝒚^(t)ϵ+2kt=1Tα(t)δ(t)ϵ+2kγ,\sum_{t=1}^{T}\alpha^{(t)}\langle\bm{x}-\widehat{\bm{x}}^{(t)},\mathbf{R}\widehat{\bm{y}}^{(t)}\rangle\leq\epsilon+2k\sum_{t=1}^{T}\alpha^{(t)}\delta^{(t)}\leq\epsilon+2k\gamma,

and the same bound applies for t=1Tα(t)𝒚𝒚^(t),𝐂𝒙^(t)\sum_{t=1}^{T}\alpha^{(t)}\langle\bm{y}-\widehat{\bm{y}}^{(t)},\mathbf{C}^{\top}\widehat{\bm{x}}^{(t)}\rangle. This shows that 𝝁^\widehat{\bm{\mu}} is an (ϵ+2kγ)(\epsilon+2k\gamma)-CCE, as claimed. This completes the proof. ∎

Lemma 3.5 thus allows us to transform a near-optimal CCE into one solely supported on the n×nn\times n submatrix corresponding to 𝐀\mathbf{A}, while incurring only a small degradation in the equilibrium quality—so long as γ\gamma is sufficiently small. Assuming now a CCE with such a property, the next lemma shows that for each individual product distribution comprising the underlying CCE, no action can be played with too high probability—at least when the corresponding weight α(t)\alpha^{(t)} is sufficiently large. Otherwise, when kk is large, a deviating player could benefit by transitioning to the part of the game where that player’s utility is k2𝐈n\frac{k}{2}\mathbf{I}_{n}. This is made precise in the following lemma.

Lemma 3.6.

Consider a TT-sparse ϵ\epsilon^{\prime}-CCE 𝛍t=1Tα(t)(𝐱(t)𝐲(t))\bm{\mu}\coloneqq\sum_{t=1}^{T}\alpha^{(t)}(\bm{x}^{(t)}\otimes\bm{y}^{(t)}) of (𝐑,𝐂)(\mathbf{R},\mathbf{C}) such that each distribution 𝐱(t)𝐲(t)\bm{x}^{(t)}\otimes\bm{y}^{(t)} is supported only on [n]×[n][n]\times[n]. Then, it holds that 𝐱i(t),𝐲j(t)(1+γ+2ϵ)/(kα(t))\bm{x}^{(t)}_{i},\bm{y}_{j}^{(t)}\leq(1+\gamma+2\epsilon^{\prime})/(k\alpha^{(t)}) for all t[T]t\in[T] and i,j[n]i,j\in[n].

An important aspect of the lemma above is that one can obtain meaningful information about each product distribution comprising 𝝁\bm{\mu} using only properties of 𝝁\bm{\mu}. Assuming for simplicity that α(1)==α(T)\alpha^{(1)}=\dots=\alpha^{(T)}, we see that the upper bound on the probabilities in Lemma 3.6 grows roughly as T/kT/k—that is, it quickly deteriorates when T1T\gg 1.

Proof of Lemma 3.6.

Fix some i[n]i\in[n]. We see that a deviation from Player yy to the n+in+i column would give that player a utility of 𝐂i,n+i(t=1Tα(t)𝒙i(t))\mathbf{C}_{i,n+i}\left(\sum_{t=1}^{T}\alpha^{(t)}\bm{x}^{(t)}_{i}\right). But the current utility of Player yy under 𝝁\bm{\mu} is at most 12(1+γ)\frac{1}{2}(1+\gamma) since 𝝁\bm{\mu} is assumed to be supported only on [n]×[n][n]\times[n]. As a result, since 𝝁\bm{\mu} is an ϵ\epsilon^{\prime}-CCE, this implies that kt=1Tα(t)𝒙i(t)(1+γ)+2ϵk\sum_{t=1}^{T}\alpha^{(t)}\bm{x}_{i}^{(t)}\leq(1+\gamma)+2\epsilon^{\prime}, which in turn yields

maxtTα(t)𝒙i(t)1+γ+2ϵk.\max_{t\leq T}\alpha^{(t)}\bm{x}_{i}^{(t)}\leq\frac{1+\gamma+2\epsilon^{\prime}}{k}.

Similar reasoning gives the same bound on maxtTα(t)𝒚j(t)\max_{t\leq T}\alpha^{(t)}\bm{y}_{j}^{(t)} for any j[n]j\in[n]. ∎

Finally, the next lemma shows how one can extract a clique on GG using a product distribution with a suitable lower bound on its welfare (which follows from Lemma 3.4) and an upper bound on the probability of its action (as in Lemma 3.6).

Lemma 3.7.

Consider a product distribution 𝐱𝐲\bm{x}\otimes\bm{y} supported only on [n]×[n][n]\times[n] with welfare at least 1+γ/6γ2T21+\gamma/\ell-6\gamma^{2}T^{2} and 𝐱i,𝐲j(1+6kγ)/\bm{x}_{i},\bm{y}_{j}\leq(1+6k\gamma)/\ell for any i,j[n]i,j\in[n], where α(t)k\ell\coloneqq\alpha^{(t)}k with α(t)1T\alpha^{(t)}\geq\frac{1}{T}. If γ<14024(+1)2k2\gamma<\frac{1}{40^{2}\ell^{4}(\ell+1)^{2}k^{2}} and k2Tk\geq 2T, the \ell largest coordinates of 𝐱\bm{x} induce an \ell-clique on GG.

Before we proceed with the proof, we will make use of a simple claim stated and proven below.

Claim 3.8.

Let 𝐱n\bm{x}\in{\mathbb{R}}^{n} be such that 0𝐱i1/0\leq\bm{x}_{i}\leq 1/\ell for all i[n]i\in[n] and i=1n𝐱i=p(0,n/]\sum_{i=1}^{n}\bm{x}_{i}=p\in(0,n/\ell]. Then, i=1n𝐱i2p/\sum_{i=1}^{n}\bm{x}_{i}^{2}\leq p/\ell.

Proof.

Let 𝒙n\bm{x}\in{\mathbb{R}}^{n} be a maximizer of i=1n𝒙i2\sum_{i=1}^{n}\bm{x}_{i}^{2} subject to the constraints specified above. Let S[n]S\subseteq[n] be the nonzero indexes of 𝒙\bm{x}. For the sake of contradiction, suppose that there exist i,iSi,i^{\prime}\in S with iii\neq i^{\prime} such that 𝒙i>𝒙i\bm{x}_{i}>\bm{x}_{i^{\prime}} and 𝒙i<1/\bm{x}_{i}<1/\ell. Then, there is a sufficiently small ϵ>0\epsilon>0 such that 𝒙i+ϵ1/\bm{x}_{i}+\epsilon\leq 1/\ell and 𝒙iϵ0\bm{x}_{i^{\prime}}-\epsilon\geq 0, and at the same time (𝒙i+ϵ)2+(𝒙iϵ)2>𝒙i2+𝒙i2(\bm{x}_{i}+\epsilon)^{2}+(\bm{x}_{i^{\prime}}-\epsilon)^{2}>\bm{x}_{i}^{2}+\bm{x}_{i^{\prime}}^{2}. This contradicts the optimality of 𝒙\bm{x}. We conclude that there are two possible cases. First, if 𝒙i=p/|S|\bm{x}_{i}=p/|S| for all iSi\in S, we have iS𝒙i2=p2/|S|p/\sum_{i\in S}\bm{x}_{i}^{2}=p^{2}/|S|\leq p/\ell (since p/|S|1/p/|S|\leq 1/\ell by feasibility). Otherwise, there is a coordinate iSi\in S such that 𝒙i<1/\bm{x}_{i}<1/\ell while 𝒙i=1/\bm{x}_{i^{\prime}}=1/\ell for iS{i}i^{\prime}\in S\setminus\{i\}. Thus, iS𝒙i2(|S|1)/2+(p(|S|1)/)2<(|S|1)/2+(p(|S|1)/)/=p/\sum_{i\in S}\bm{x}_{i}^{2}\leq(|S|-1)/\ell^{2}+(p-(|S|-1)/\ell)^{2}<(|S|-1)/\ell^{2}+(p-(|S|-1)/\ell)/\ell=p/\ell. ∎

Proof of Lemma 3.7.

Given that 𝒙𝒚\bm{x}\otimes\bm{y} is supported only on [n]×[n][n]\times[n], we have

(1+γ)i=1n𝒙i𝒚i+(1i=1n𝒙i𝒚i)𝖲𝖶(𝒙𝒚)1+γ/6γ2T2,(1+\gamma)\sum_{i=1}^{n}\bm{x}_{i}\bm{y}_{i}+\left(1-\sum_{i=1}^{n}\bm{x}_{i}\bm{y}_{i}\right)\geq\operatorname{\mathsf{SW}}(\bm{x}\otimes\bm{y})\geq 1+\gamma/\ell-6\gamma^{2}T^{2},

which implies that i=1n𝒙i𝒚i1/6γT2\sum_{i=1}^{n}\bm{x}_{i}\bm{y}_{i}\geq 1/\ell-6\gamma T^{2}. Further, 3.8 implies that i=1n𝒙i2(1+6kγ)/\sum_{i=1}^{n}\bm{x}_{i}^{2}\leq(1+6k\gamma)/\ell and i=1n𝒚i2(1+6kγ)/\sum_{i=1}^{n}\bm{y}_{i}^{2}\leq(1+6k\gamma)/\ell. Thus, combining the previous inequalities,

𝒙𝒚2=i=1n𝒙i2+i=1n𝒚i22i=1n𝒙i𝒚i12kγ+12γT215k2γ.\|\bm{x}-\bm{y}\|^{2}=\sum_{i=1}^{n}\bm{x}_{i}^{2}+\sum_{i=1}^{n}\bm{y}_{i}^{2}-2\sum_{i=1}^{n}\bm{x}_{i}\bm{y}_{i}\leq\frac{12k\gamma+12\gamma T^{2}}{\ell}\leq\frac{15k^{2}\gamma}{\ell}. (7)

Moreover,

1𝒙,𝒚=𝒙,𝒚𝒙+𝒙2\displaystyle\frac{1}{\ell}\leq\langle\bm{x},\bm{y}\rangle=\langle\bm{x},\bm{y}-\bm{x}\rangle+\|\bm{x}\|^{2} 𝒙𝒙𝒚+𝒙2\displaystyle\leq\|\bm{x}\|\|\bm{x}-\bm{y}\|+\|\bm{x}\|^{2}
1+6kγ15k2γ+𝒙2\displaystyle\leq\sqrt{\frac{1+6k\gamma}{\ell}}\sqrt{\frac{15k^{2}\gamma}{\ell}}+\|\bm{x}\|^{2}
4kγ+𝒙2\displaystyle\leq\frac{4k\sqrt{\gamma}}{\ell}+\|\bm{x}\|^{2}

since 1+4kγ16/15\sqrt{1+4k\gamma}\leq\sqrt{16/15}. Rearranging,

𝒙21(14kγ).\|\bm{x}\|^{2}\geq\frac{1}{\ell}\left(1-4k\sqrt{\gamma}\right). (8)

Now, let S{i[n]:𝒙i(1γ)/}S\coloneqq\{i\in[n]:\bm{x}_{i}\geq(1-\gamma^{\prime})/\ell\} for γ402kγ\gamma^{\prime}\coloneqq 40\ell^{2}k\sqrt{\gamma}, and piS𝒙ip\coloneqq\sum_{i\in S}\bm{x}_{i}. Applying 3.8 to both SS and [n]S[n]\setminus S,

𝒙2=iS𝒙i2+i[n]S𝒙i2p1+6kγ+(1p)1γ.\|\bm{x}\|^{2}=\sum_{i\in S}\bm{x}_{i}^{2}+\sum_{i\in[n]\setminus S}\bm{x}_{i}^{2}\leq p\frac{1+6k\gamma}{\ell}+(1-p)\frac{1-\gamma^{\prime}}{\ell}.

Combining with (8), we get

pγ4kγ6kγ+γ1142,p\geq\frac{\gamma^{\prime}-4k\sqrt{\gamma}}{6k\gamma+\gamma^{\prime}}\geq 1-\frac{1}{4\ell^{2}},

where we used that γ=402kγ\gamma^{\prime}=40\ell^{2}k\sqrt{\gamma}. We now claim that |S|=|S|=\ell. Indeed, since (1γ)/𝒙i(1+6kγ)/(1-\gamma^{\prime})/\ell\leq\bm{x}_{i}\leq(1+6k\gamma)/\ell for all iSi\in S, we have

1γ|S|iS𝒙i1|S|1γ=1402kγ<+1\frac{1-\gamma^{\prime}}{\ell}|S|\leq\sum_{i\in S}\bm{x}_{i}\leq 1\implies|S|\leq\frac{\ell}{1-\gamma^{\prime}}=\frac{\ell}{1-40\ell^{2}k\sqrt{\gamma}}<\ell+1

for γ<14024(+1)2k2\gamma<\frac{1}{40^{2}\ell^{4}(\ell+1)^{2}k^{2}}, and (since p=iS𝒙i11/(42)p=\sum_{i\in S}\bm{x}_{i}\geq 1-1/(4\ell^{2}))

1142iS𝒙i|S|1+6kγ|S|>1.1-\frac{1}{4\ell^{2}}\leq\sum_{i\in S}\bm{x}_{i}\leq|S|\frac{1+6k\gamma}{\ell}\implies|S|>\ell-1.

Thus, |S|=|S|=\ell. Finally, we show that SS is an \ell-clique on GG. We first note that, by (7),

𝒚j𝒙j15k2γ1γ4kγ11.5(+1)12\bm{y}_{j}\geq\bm{x}_{j}-\sqrt{\frac{15k^{2}\gamma}{\ell}}\geq\frac{1-\gamma^{\prime}}{\ell}-4k\sqrt{\frac{\gamma}{\ell}}\geq\frac{1}{\ell}-\frac{1.5}{\ell(\ell+1)}\geq\frac{1}{2\ell}

for all jSj\in S, where the penultimate inequality follows from our choice of γ\gamma and γ\gamma^{\prime}, and the last inequality uses that 2\ell\geq 2. If we assume for the sake of contradiction that there exist i,jSi^{\prime},j^{\prime}\in S such that 𝐀i,j=0\mathbf{A}_{i^{\prime},j^{\prime}}=0, that is, SS is not a clique on GG, then

𝖲𝖶(𝒙𝒚)\displaystyle\operatorname{\mathsf{SW}}(\bm{x}\otimes\bm{y}) =i=1nj=1n𝒙i(𝐀i,j+γ𝟙{i=j})𝒚j\displaystyle=\sum_{i=1}^{n}\sum_{j=1}^{n}\bm{x}_{i}(\mathbf{A}_{i,j}+\gamma\mathbbm{1}\{i=j\})\bm{y}_{j}
(1+γ)i=1n𝒙i𝒚i+i=1nji𝟙{{i,j}{i,j}}𝒙i𝐀i,j𝒚j+2𝒙i𝐀i,j𝒚j\displaystyle\leq(1+\gamma)\sum_{i=1}^{n}\bm{x}_{i}\bm{y}_{i}+\sum_{i=1}^{n}\sum_{j\neq i}\mathbbm{1}\{\{i,j\}\neq\{i^{\prime},j^{\prime}\}\}\bm{x}_{i}\mathbf{A}_{i,j}\bm{y}_{j}+2\bm{x}_{i^{\prime}}\mathbf{A}_{i^{\prime},j^{\prime}}\bm{y}_{j^{\prime}}
(1+γ)1+6kγ+11+6kγ122\displaystyle\leq(1+\gamma)\frac{1+6k\gamma}{\ell}+1-\frac{1+6k\gamma}{\ell}-\frac{1}{2\ell^{2}}
1+γ12<1+γ6γ2T2,\displaystyle\leq 1+\frac{\gamma}{\ell}-\frac{1}{\ell^{2}}<1+\frac{\gamma}{\ell}-6\gamma^{2}T^{2},

where the last inequality follows from our choice of γ\gamma. This is a contradiction since 𝖲𝖶(𝒙𝒚)1+γ/6γ2T2\operatorname{\mathsf{SW}}(\bm{x}\otimes\bm{y})\geq 1+\gamma/\ell-6\gamma^{2}T^{2}. ∎

We are now ready to combine Lemmas 3.4, 3.5, 3.6 and 3.7 in order to prove Theorem 3.1.

Proof of Theorem 3.1.

It suffices to show that if GG contains a kk-clique, then Algorithm 1 outputs a k/Tk/T-clique. Indeed, if ω(G)2T\omega(G)\leq 2T, then Algorithm 1 trivially returns a 2T2T-approximation; otherwise, it suffices to consider the largest iteration of the algorithm such that kω(G)k\leq\omega(G). Consider now for that iteration the TT-sparse ϵ\epsilon-CCE, 𝝁=t=1Tα(t)(𝒙(t)𝒚(t))\bm{\mu}=\sum_{t=1}^{T}\alpha^{(t)}(\bm{x}^{(t)}\otimes\bm{y}^{(t)}), given as output by OptimalSparseCCE in Algorithm 1; we can assume here that α(t)>0\alpha^{(t)}>0 for all t[T]t\in[T]. Let 𝝁^t=1Tα(t)(𝒙^(t)𝒚^(t))\widehat{\bm{\mu}}\coloneqq\sum_{t=1}^{T}\alpha^{(t)}(\widehat{\bm{x}}^{(t)}\otimes\widehat{\bm{y}}^{(t)}), where (𝒙^(t))tT(\widehat{\bm{x}}^{(t)})_{t\leq T} and (𝒚^(t))tT(\widehat{\bm{y}}^{(t)})_{t\leq T} are as in Algorithm 1. By Lemma 3.4, we know that there is a TT-sparse CCE with welfare at least 1+γT/k1+\gamma T/k, and so since the optimality gap ϵ^\hat{\epsilon} is small enough (by Algorithm 1), we get that 𝖲𝖶(𝝁)1\operatorname{\mathsf{SW}}(\bm{\mu})\geq 1. Thus, we can apply Lemma 3.5 to conclude that 𝝁^\widehat{\bm{\mu}} is an (ϵ+2kγ)(\epsilon+2k\gamma)-CCE, which is a (5kγ/2)(5k\gamma/2)-CCE by our choice of ϵ\epsilon (Algorithm 1). Next, Lemma 3.6 implies that 𝒙^i(t)(1+γ+5kγ)/(α(t)k)\widehat{\bm{x}}_{i}^{(t)}\leq(1+\gamma+5k\gamma)/(\alpha^{(t)}k) for all i[n]i\in[n] and t[T]t\in[T]. Using 3.8 and Cauchy-Schwarz, we have

i=1n𝒙^i(t)𝒚^i(t)(i=1n(𝒙^i(t))2)(i=1n(𝒚^i(t))2)1+γ+5kγα(t)k.\sum_{i=1}^{n}\widehat{\bm{x}}_{i}^{(t)}\widehat{\bm{y}}_{i}^{(t)}\leq\sqrt{\left(\sum_{i=1}^{n}(\widehat{\bm{x}}_{i}^{(t)})^{2}\right)\left(\sum_{i=1}^{n}(\widehat{\bm{y}}_{i}^{(t)})^{2}\right)}\leq\frac{1+\gamma+5k\gamma}{\alpha^{(t)}k}.

Thus, for all t[T]t\in[T],

𝖲𝖶(𝒙^(t)𝒚^(t))1+γ+γ2+5kγ2α(t)k.\operatorname{\mathsf{SW}}(\widehat{\bm{x}}^{(t)}\otimes\widehat{\bm{y}}^{(t)})\leq 1+\frac{\gamma+\gamma^{2}+5k\gamma^{2}}{\alpha^{(t)}k}. (9)

By Lemmas 3.4 and 3.5, we have

α(t)𝖲𝖶(𝒙^(t)𝒚^(t))\displaystyle\alpha^{(t)}\operatorname{\mathsf{SW}}(\widehat{\bm{x}}^{(t)}\otimes\widehat{\bm{y}}^{(t)}) 1+γTkϵ^τtα(τ)𝖲𝖶(𝒙^(τ)𝒚^(τ))\displaystyle\geq 1+\frac{\gamma T}{k}-\hat{\epsilon}-\sum_{\tau\neq t}\alpha^{(\tau)}\operatorname{\mathsf{SW}}(\widehat{\bm{x}}^{(\tau)}\otimes\widehat{\bm{y}}^{(\tau)})
α(t)+γTkϵ^(T1)γ+γ2+5kγ2k\displaystyle\geq\alpha^{(t)}+\frac{\gamma T}{k}-\hat{\epsilon}-(T-1)\frac{\gamma+\gamma^{2}+5k\gamma^{2}}{k} (10)
α(t)+γk5γ2TTkγ2ϵ^α(t)+γk6γ2T,\displaystyle\geq\alpha^{(t)}+\frac{\gamma}{k}-5\gamma^{2}T-\frac{T}{k}\gamma^{2}-\hat{\epsilon}\geq\alpha^{(t)}+\frac{\gamma}{k}-6\gamma^{2}T, (11)

where (10) uses (9) and the fact that α(t)=1τtα(τ)\alpha^{(t)}=1-\sum_{\tau\neq t}\alpha^{(\tau)}, and (11) follows from our choice of ϵ^\hat{\epsilon} and the fact that k2k\geq 2. Now, let t[T]t^{\star}\in[T] be such that α(t)1/T\alpha^{(t^{\star})}\geq 1/T; Algorithm 1 returns such a tt^{\star}. Then, (11) implies that 𝖲𝖶(𝒙^(t)𝒚^(t))1+γα(t)k6γ2T2\operatorname{\mathsf{SW}}(\widehat{\bm{x}}^{(t^{\star})}\otimes\widehat{\bm{y}}^{(t^{\star})})\geq 1+\frac{\gamma}{\alpha^{(t^{\star})}k}-6\gamma^{2}T^{2} since α(t)1/T\alpha^{(t^{\star})}\geq 1/T. Finally, we can use Lemma 3.7 with respect to 𝒙^(t)𝒚^(t)\widehat{\bm{x}}^{(t^{\star})}\otimes\widehat{\bm{y}}^{(t^{\star})} to obtain an (α(t)k)(\alpha^{(t^{\star})}k)-clique, and thereby a k/Tk/T-clique in GG since α(t)1/T\alpha^{(t^{\star})}\geq 1/T. This completes the proof. ∎

3.2 Further implications

Building on our previous reduction, we first obtain some further lower bounds for natural problems related to sparse CCE, extending some corresponding results of [GZ89]. The reduction here serves as a warm-up for the upcoming one in Section 3.3, which will enable us to capture other natural objectives, beyond welfare, and preclude any multiplicative approximation.

In this context, the problem UniqueUniformSparseCCE(𝒢,T)\textsc{UniqueUniformSparseCCE}({\mathcal{G}},T) asks whether the two-player game 𝒢{\mathcal{G}} given as input admits a unique uniform TT-sparse CCE (recall Definition 2.1); no approximation error is allowed here, for otherwise there are trivially multiple sparse CCE. The reason why we need to posit a uniform TT-sparse distribution is explained more after the proof of Theorem 3.9. We further point out that UniqueUniformSparseCCE does not hinge on an underlying objective. As before, it is easy to see that without the sparsity constraint there is a polynomial-time algorithm based on linear programming. We will show that obtaining strongly sublinear sparsity is again computationally hard (Theorem 3.9).

The basic idea here is to augment the game given in (3) with an additional action for each player, say 𝖮\mathsf{O}, such that the following property holds: if the resulting game has a unique uniform TT-sparse CCE, namely (𝖮,𝖮)(\mathsf{O},\mathsf{O}), then GG does not contain a clique of size kk; otherwise, GG must contain a clique of size Ω(k/T)\Omega(k/T). Relying then on the hardness result of [Zuc07], we arrive at the following theorem.

1 2 3 4 𝖮\mathsf{O}
1 (1+γ,1+γ)(1+\gamma,1+\gamma) (1, 1) (1, 1) (1, 1) (k,k)(-k,k) (0, 0) (0, 0) (0, 0) (2r,2r)(-2r,2r)
2 (1, 1) (1+γ,1+γ)(1+\gamma,1+\gamma) (1, 1) (0, 0) 0 (k,k)(-k,k) (0, 0) (0, 0) (2r,2r)(-2r,2r)
3 (1, 1) (1, 1) (1+γ,1+γ)(1+\gamma,1+\gamma) (1,1) (0, 0) (0, 0) (k,k)(-k,k) (0, 0) (2r,2r)(-2r,2r)
4 (1, 1) (0, 0) (1, 1) (1+γ,1+γ)(1+\gamma,1+\gamma) (0, 0) (0, 0) (0, 0) (k,k)(-k,k) (2r,2r)(-2r,2r)
(k,k)(k,-k) (0, 0) (0, 0) (0, 0) (0, 0) (0, 0) (0, 0) (0, 0) (2r,2r)(-2r,2r)
(0, 0) (k,k)(k,-k) (0, 0) (0, 0) (0, 0) (0, 0) (0, 0) (0, 0) (2r,2r)(-2r,2r)
(0, 0) (0, 0) (k,k)(k,-k) (0, 0) (0, 0) (0, 0) (0, 0) (0, 0) (2r,2r)(-2r,2r)
(0, 0) (0, 0) (0, 0) (k,k)(k,-k) (0, 0) (0, 0) (0, 0) (0, 0) (2r,2r)(-2r,2r)
𝖮\mathsf{O} (2r,2r)(2r,-2r) (2r,2r)(2r,-2r) (2r,2r)(2r,-2r) (2r,2r)(2r,-2r) (2r,2r)(2r,-2r) (2r,2r)(2r,-2r) (2r,2r)(2r,-2r) (2r,2r)(2r,-2r) (2r,2r)(2r,2r)
Figure 3: Payoff matrices of 𝒢{\mathcal{G}}^{\prime} (after multiplying each entry by 22) based on graph GG of Figure 2.
Theorem 3.9.

UniqueUniformSparseCCE(𝒢,n1ϵ)\textsc{UniqueUniformSparseCCE}({\mathcal{G}},n^{1-\epsilon}) is \NP\NP-hard with respect to n×nn\times n games for any constant ϵ>0\epsilon>0.

Proof.

We will use the fact that distinguishing whether an nn-node graph GG has a clique of size at most nϵn^{\epsilon} or at least n1ϵn^{1-\epsilon} is \NP-hard for any constant ϵ>0\epsilon>0 [Zuc07]. We proceed as follows. We augment the game given in (3) with an additional action, 𝖮\mathsf{O}, for each player. We will refer to the resulting game as 𝒢{\mathcal{G}}^{\prime}. When both players select 𝖮\mathsf{O}, they both receive a utility of r(1+γT/k)/2r\coloneqq(1+\gamma T/k)/2. When only one player selects 𝖮\mathsf{O}, that player receives rr while the other player receives r-r (see Figure 3). It is clear that (𝖮,𝖮)(\mathsf{O},\mathsf{O}) is a (pure) Nash equilibrium, and hence a TT-sparse CCE. Moreover, Lemma 3.4 shows that when GG contains a kk-clique, there is a uniform TT-sparse correlated distribution 𝝁\bm{\mu} in 𝒢{\mathcal{G}} supported only on the n×nn\times n submatrix corresponding to 𝐀\mathbf{A}, in which each player obtains a utility of rr. It is easy to see that 𝝁\bm{\mu} induces a (uniform) TT-sparse CCE in 𝒢{\mathcal{G}}^{\prime} as well. What remains to show thus is that when GG does not contain an Ω(k/T)\Omega(k/T) clique, (𝖮,𝖮)(\mathsf{O},\mathsf{O}) is the only uniform TT-sparse CCE of 𝒢{\mathcal{G}}^{\prime}.

Indeed, since a player can always switch to playing 𝖮\mathsf{O}, which secures that player a utility of rr, any TT-sparse CCE in 𝒢{\mathcal{G}}^{\prime} must have welfare at least 2r=1+γT/k2r=1+\gamma T/k. Following the argument of Section 3.1.1, when γ\gamma is sufficiently small, we will show that this can only happen under (𝖮,𝖮)(\mathsf{O},\mathsf{O}) or when GG contains an Ω(k/T)\Omega(k/T)-clique. In proof, let 𝝁=1Tt=1T𝒙(t)𝒚(t)\bm{\mu}=\frac{1}{T}\sum_{t=1}^{T}\bm{x}^{(t)}\otimes\bm{y}^{(t)} be a CCE in 𝒢{\mathcal{G}}^{\prime} distinct from (𝖮,𝖮)(\mathsf{O},\mathsf{O}), which we argued must satisfy 𝖲𝖶(𝝁)1+γT/k\operatorname{\mathsf{SW}}(\bm{\mu})\geq 1+\gamma T/k. We consider now the product t[T]t\in[T] that maximizes 𝖲𝖶(𝒙(t)𝒚(t))\operatorname{\mathsf{SW}}(\bm{x}^{(t)}\otimes\bm{y}^{(t)}) subject to the constraint that the probability mass assigned to the n×nn\times n submatrix corresponding to 𝐀\mathbf{A} is (strictly) positive. It is clear that such tt exists, and 𝖲𝖶(𝒙(t)𝒚(t))1+γT/k\operatorname{\mathsf{SW}}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})\geq 1+\gamma T/k. We will now argue that pi𝒙(t),j𝒚(t)[in,jn]12γp\coloneqq\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i\leq n,j\leq n]\geq 1-2\gamma. Let p𝖮=i𝒙(t),j𝒚(t)[i=𝖮,j=𝖮]p_{\mathsf{O}}=\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i=\mathsf{O},j=\mathsf{O}]. Then, it follows that 𝖲𝖶(𝒙(t)𝒚(t))p(1+γ)+p𝖮(1+γT/k)\operatorname{\mathsf{SW}}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})\leq p(1+\gamma)+p_{\mathsf{O}}(1+\gamma T/k) since 𝒢{\mathcal{G}}^{\prime} is, by construction, zero-sum when excluding (𝖮,𝖮)(\mathsf{O},\mathsf{O}) and the n×nn\times n submatrix corresponding to 𝐀\mathbf{A}. Thus, p+p𝖮1γp+p_{\mathsf{O}}\geq 1-\gamma. Further, a union bound implies that p+p𝖮+2pp𝖮1p+p_{\mathsf{O}}+2pp_{\mathsf{O}}\leq 1. Indeed, we have

i𝒙(t),j𝒚(t)[i=𝖮,j𝖮]\displaystyle\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i=\mathsf{O},j\neq\mathsf{O}] =i𝒙(t)[i=𝖮]j𝒚(t)[j𝖮]\displaystyle=\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)}}[i=\mathsf{O}]\operatorname{\mathbb{P}}_{j\sim\bm{y}^{(t)}}[j\neq\mathsf{O}]
i𝒙(t),j𝒚(t)[i=𝖮,j=𝖮]i𝒙(t),j𝒚(t)[in,j𝖮]\displaystyle\geq\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i=\mathsf{O},j=\mathsf{O}]\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i\leq n,j\neq\mathsf{O}]
p𝖮p,\displaystyle\geq p_{\mathsf{O}}p,

and the same lower bound applies to i𝒙(t),j𝒚(t)[i𝖮,j=𝖮]\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i\neq\mathsf{O},j=\mathsf{O}]. Combining with the fact that 𝖲𝖶(𝒙(t)𝒚(t))1+γT/k\operatorname{\mathsf{SW}}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})\geq 1+\gamma T/k, we conclude that

1+γTk12pp𝖮+pγ+p𝖮γTk1+γTk2pp𝖮+pγp𝖮γ2,1+\gamma\frac{T}{k}\leq 1-2pp_{\mathsf{O}}+p\gamma+p_{\mathsf{O}}\gamma\frac{T}{k}\leq 1+\gamma\frac{T}{k}-2pp_{\mathsf{O}}+p\gamma\implies p_{\mathsf{O}}\leq\frac{\gamma}{2},

where we used that p>0p>0 (by assumption of how t[T]t\in[T] was chosen). Next, we define 𝒙^in(t)=𝒙in(t)/i=1n𝒙i(t)\widehat{\bm{x}}_{i\leq n}^{(t)}=\nicefrac{{\bm{x}^{(t)}_{i\leq n}}}{{\sum_{i=1}^{n}\bm{x}_{i}^{(t)}}} and 𝒚^jn(t)=𝒚jn(t)/j=1n𝒚j(t)\widehat{\bm{y}}_{j\leq n}^{(t)}=\nicefrac{{\bm{y}^{(t)}_{j\leq n}}}{{\sum_{j=1}^{n}\bm{y}_{j}^{(t)}}}. Similarly to Lemma 3.6, we will now show that 𝒙^i(t),𝒚^j(t)2/\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}_{j}^{(t)}\leq 2/\ell for any (i,j)[n]×[n](i,j)\in[n]\times[n], where k/T\ell\coloneqq k/T and γ\gamma is sufficiently small. Indeed, by construction of 𝒢{\mathcal{G}}^{\prime},

1+γ𝖲𝖶(𝝁)(1+γ)(i,j)𝝁[(in,jn)(i=𝖮,j=𝖮)].1+\frac{\gamma}{\ell}\leq\operatorname{\mathsf{SW}}(\bm{\mu})\leq(1+\gamma)\operatorname{\mathbb{P}}_{(i,j)\sim\bm{\mu}}[(i\leq n,j\leq n)\lor(i=\mathsf{O},j=\mathsf{O})].

Thus,

(i,j)𝝁[(in,jn)(i=𝖮,j=𝖮)]1γ.\operatorname{\mathbb{P}}_{(i,j)\sim\bm{\mu}}[(i\leq n,j\leq n)\lor(i=\mathsf{O},j=\mathsf{O})]\geq 1-\gamma.

This implies that the expected utility of each player under 𝝁\bm{\mu} is at most (1+γ+kγ)/2(1+\gamma+k\gamma)/2. Further, since 𝝁\bm{\mu} is also assumed to be a CCE, considering the deviation of Player yy to column n+in+i, we have that for any i[n]i\in[n],

1+γ+kγ2k2T𝒙i(t)T(1γ)Tr𝒙^i(t)2\frac{1+\gamma+k\gamma}{2}\geq\frac{k}{2T}\bm{x}_{i}^{(t)}-\frac{T-(1-\gamma)}{T}r\implies\widehat{\bm{x}}_{i}^{(t)}\leq\frac{2}{\ell}

when γ\gamma is sufficiently small, where we used the fact that i=1n𝒙i(t)p1γ\sum_{i=1}^{n}\bm{x}_{i}^{(t)}\geq p\geq 1-\gamma. Similar reasoning yields that 𝒚^j(t)2/\widehat{\bm{y}}_{j}^{(t)}\leq 2/\ell for any j[n]j\in[n]. Moreover, since 𝖲𝖶(𝒙(t)𝒚(t))1+γ/\operatorname{\mathsf{SW}}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})\geq 1+\gamma/\ell, it is easy to see that 𝖲𝖶(𝒙^(t)𝒚^(t))1+γ/\operatorname{\mathsf{SW}}(\widehat{\bm{x}}^{(t)}\otimes\widehat{\bm{y}}^{(t)})\geq 1+\gamma/\ell, which in turn implies that i=1n𝒙^i(t)𝒚^i(t)1/\sum_{i=1}^{n}\widehat{\bm{x}}_{i}^{(t)}\widehat{\bm{y}}_{i}^{(t)}\geq 1/\ell. To conclude the argument, we will show—analogously to Lemma 3.7—how to extract an Ω()\Omega(\ell)-clique from that product distribution. Namely, we identify the set SS consisting of all i[n]i\in[n] such that 𝒙^i(t),𝒚^i(t)1/(16)\widehat{\bm{x}}^{(t)}_{i},\widehat{\bm{y}}^{(t)}_{i}\geq 1/(16\ell). Then,

1i=1n𝒙^i(t)𝒚^i(t)\displaystyle\frac{1}{\ell}\leq\sum_{i=1}^{n}\widehat{\bm{x}}_{i}^{(t)}\widehat{\bm{y}}_{i}^{(t)} =iS𝒙^i(t)𝒚^i(t)+iS𝒙^i(t)𝒚^i(t)\displaystyle=\sum_{i\in S}\widehat{\bm{x}}_{i}^{(t)}\widehat{\bm{y}}_{i}^{(t)}+\sum_{i\notin S}\widehat{\bm{x}}_{i}^{(t)}\widehat{\bm{y}}_{i}^{(t)}
=iS𝒙^i(t)𝒚^i(t)+iSmax(𝒙^i(t),𝒚^i(t))min(𝒙^i(t),𝒚^i(t))\displaystyle=\sum_{i\in S}\widehat{\bm{x}}_{i}^{(t)}\widehat{\bm{y}}_{i}^{(t)}+\sum_{i\notin S}\max(\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}_{i}^{(t)})\min(\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}_{i}^{(t)})
|S|42+iSmax(𝒙^i(t),𝒚^i(t))2iSmin(𝒙^i(t),𝒚^i(t))2\displaystyle\leq|S|\frac{4}{\ell^{2}}+\sqrt{\sum_{i\notin S}\max(\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}_{i}^{(t)})^{2}\sum_{i\notin S}\min(\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}_{i}^{(t)})^{2}} (12)
|S|42+4116,\displaystyle\leq|S|\frac{4}{\ell^{2}}+\sqrt{\frac{4}{\ell}\frac{1}{16\ell}}, (13)

where (12) uses the Cauchy-Schwarz inequality and the fact that 𝒙^i(t),𝒚^i(t)2/\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}^{(t)}_{i}\leq 2/\ell, and (13) is a consequence of 3.8; to apply 3.8, we note that iSmax(𝒙^i(t),𝒚^i(t))iS(𝒙^i(t)+𝒚^i(t))2\sum_{i\notin S}\max(\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}_{i}^{(t)})\leq\sum_{i\notin S}(\widehat{\bm{x}}_{i}^{(t)}+\widehat{\bm{y}}_{i}^{(t)})\leq 2; iSmin(𝒙^i(t),𝒚^i(t))iS𝒙^i(t)1\sum_{i\notin S}\min(\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}_{i}^{(t)})\leq\sum_{i\notin S}\widehat{\bm{x}}_{i}^{(t)}\leq 1; max(𝒙^i(t),𝒚^i(t))2/\max(\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}_{i}^{(t)})\leq 2/\ell for all i[n]i\in[n]; and min(𝒙^i(t),𝒚^i(t))1/(16)\min(\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}_{i}^{(t)})\leq 1/(16\ell) for all i[n]Si\in[n]\setminus S (by definition of SS). As a result, we conclude that |S|/8|S|\geq\ell/8. Finally, similarly to the argument of Lemma 3.7, it is easy to see that SS induces a clique on GG—the contrary case would contradict the fact that 𝖲𝖶(𝒙^(t)𝒚^(t))1+γ/\operatorname{\mathsf{SW}}(\widehat{\bm{x}}^{(t)}\otimes\widehat{\bm{y}}^{(t)})\geq 1+\gamma/\ell when γ\gamma is sufficiently small.

Therefore, under uniqueness we know that GG does not contain a clique of size kk, while in the contrary case GG must contain a clique of size Ω(k/T)\Omega(k/T). Taking k=n1ϵk=n^{1-\epsilon} and T=Θ(n12ϵ)T=\Theta(n^{1-2\epsilon}) in conjunction with the hardness result of [Zuc07] completes the proof. ∎

It is not clear how to extend Theorem 3.9 when we allow non-uniform mixtures. In particular, taking two mixtures on (1,1)(1,1) and (𝖮,𝖮)(\mathsf{O},\mathsf{O}) with weights α(1)0\alpha^{(1)}\approx 0 and α(2)1\alpha^{(2)}\approx 1, respectively, yields a (22-sparse) CCE for 𝒢{\mathcal{G}}^{\prime}, but clearly does not contain any useful information about GG. (Interestingly, the previously described correlated distribution is not a CE: a player can profitably deviate according to 1n+11\mapsto n+1 and 𝖮𝖮\mathsf{O}\mapsto\mathsf{O}.)

Next, the problem SubsetUniformSparseCCE(𝒢,T,S)\textsc{SubsetUniformSparseCCE}({\mathcal{G}},T,S) asks whether there is a uniform TT-sparse CCE supported solely on the joint actions given as input in SS; this problem also does not hinge on an underlying objective. Using the previous reduction, by excluding the pair (𝖮,𝖮)(\mathsf{O},\mathsf{O}), we obtain a similar \NP-hardness result.

Theorem 3.10.

SubsetUniformSparseCCE(𝒢,n1ϵ,S)\textsc{SubsetUniformSparseCCE}({\mathcal{G}},n^{1-\epsilon},S) is \NP\NP-hard with respect to n×nn\times n games for any constant ϵ>0\epsilon>0.

3.3 Inapproximability and other objectives

We will next show that a small adjustment to the reduction of Section 3.2 has two further important consequences: it enables capturing other natural objectives (Corollaries 3.13 and 3.12), beyond welfare, and it precludes any multiplicative approximation with respect to the objective (Corollary 3.14).

The driving force behind those results is Theorem 3.11 below. For simplicity, in what follows we prove lower bounds concerning CCE (that is, with ϵ=0\epsilon=0), although it is straightforward to extend the argument so as to account for some imprecision in the approximation of the equilibrium.

Theorem 3.11.

Given as input an nn-node graph GG, k[n]k\in[n] and ϵ1\epsilon\ll 1, we can construct a (2n+1)×(2n+1)(2n+1)\times(2n+1) (two-player) game with the following properties:

  1. 1.

    it always admits a (pure) Nash equilibrium under which both players obtain a utility of ϵ\epsilon;

  2. 2.

    when GG contains a clique of size kk, there is a uniform TT-sparse CCE such that each player obtains a utility of at least 1/21/2; and

  3. 3.

    when GG does not contain a clique of size k/(8T)k/(8T), there is no other uniform TT-sparse CCE.

Assuming Theorem 3.11, we immediately establish \NP-hardness with respect to other natural objectives, besides the social welfare—any reasonable objective would refrain from selecting the former equilibrium (Item 1), if there is a choice to do so. In particular, we mention below two notable such implications.

Corollary 3.12.

Computing a uniform TT-sparse CCE that maximizes the utility of Player xx (or Player yy) is \NP\NP-hard with respect to n×nn\times n games for any Tn1ϵT\leq n^{1-\epsilon} and constant ϵ>0\epsilon>0.

For the next implication, we recall that the egalitarian social welfare is the expected utility of the player who is worse—that objective is often advocated in the context of fairness.444Throughout this paper, we follow the convention that (social) welfare without any further specification refers to the utilitarian social welfare.

Corollary 3.13.

Computing a uniform TT-sparse CCE that maximizes the egalitarian social welfare is \NP\NP-hard with respect to n×nn\times n games for any Tn1ϵT\leq n^{1-\epsilon} and constant ϵ>0\epsilon>0.

Moreover, Theorem 3.11 implies hardness of approximating the (utilitarian) welfare, measured here as a ratio. In particular, to make this problem meaningful, the reduction of Theorem 3.11 makes sure that any joint action profile has nonnegative welfare.

Corollary 3.14.

Computing a uniform TT-sparse CCE that approximates the welfare-optimal one to any positive ratio is \NP-hard with respect to n×nn\times n games for any Tn1ϵT\leq n^{1-\epsilon} and constant ϵ>0\epsilon>0.

We mention in passing that Theorem 3.11 resembles to a certain extent a result of [CS08] concerning Nash equilibria. Their reduction is instead based on SAT, and it would be interesting to understand how it interacts with sparse CCE. Before we proceed with the proof of Theorem 3.11, we also point out that Theorems 3.9 and 3.10 are implied by Theorem 3.11, but the reduction in Section 3.2 is easier to work with because deviating to 𝖮\mathsf{O} secures the same utility no matter how the other player acts (Figure 3); this will no longer be the case for the reduction behind Theorem 3.11, complicating the argument.

Proof of Theorem 3.11.

The reduction is based on game 𝒢{\mathcal{G}}^{\prime} in the reduction of Theorem 3.9, with the only difference that the utility under (𝖮,𝖮)(\mathsf{O},\mathsf{O}) is now taken to be 0<ϵ10<\epsilon\ll 1 for both players. As before, (𝖮,𝖮)(\mathsf{O},\mathsf{O}) is a (pure) Nash equilibrium, and hence a TT-sparse CCE (implying Item 1). Item 2 also follows readily. It thus suffices to show that when GG does not contain a k/(8T)k/(8T)-clique, (𝖮,𝖮)(\mathsf{O},\mathsf{O}) is the only uniform TT-sparse CCE (Item 3).

Let 𝝁\bm{\mu} be a TT-sparse CCE distinct from (𝖮,𝖮)(\mathsf{O},\mathsf{O}). Considering the deviation of Player xx to 𝖮\mathsf{O} and of Player yy to 𝖮\mathsf{O}, we have

𝖲𝖶(𝝁)\displaystyle\operatorname{\mathsf{SW}}(\bm{\mu}) r(11Tt=1T𝒙𝖮(t)+11Tt=1T𝒚𝖮(t))+ϵ(1Tt=1T𝒙𝖮(t)+1Tt=1T𝒚𝖮(t))\displaystyle\geq r\left(1-\frac{1}{T}\sum_{t=1}^{T}\bm{x}_{\mathsf{O}}^{(t)}+1-\frac{1}{T}\sum_{t=1}^{T}\bm{y}_{\mathsf{O}}^{(t)}\right)+\epsilon\left(\frac{1}{T}\sum_{t=1}^{T}\bm{x}_{\mathsf{O}}^{(t)}+\frac{1}{T}\sum_{t=1}^{T}\bm{y}_{\mathsf{O}}^{(t)}\right) (14)
1Tt=1T((1+γ)(1𝒙𝖮(t)+𝒚𝖮(t)2)+2ϵ𝒙𝖮(t)+𝒚𝖮(t)2)γ1Tt=1T(1𝒙𝖮(t)+𝒚𝖮(t)2).\displaystyle\geq\frac{1}{T}\sum_{t=1}^{T}\left((1+\gamma)\left(1-\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\right)+2\epsilon\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\right)-\gamma\frac{1}{T}\sum_{t=1}^{T}\left(1-\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\right). (15)

Further, since the game is zero-sum when excluding (𝖮,𝖮)(\mathsf{O},\mathsf{O}) and the joint actions in the n×nn\times n submatrix corresponding to 𝐀\mathbf{A},

𝖲𝖶(𝝁)\displaystyle\operatorname{\mathsf{SW}}(\bm{\mu}) 2ϵ1Tt=1T𝒙𝖮(t)𝒚𝖮(t)+(1+γ)1Tt=1Ti𝒙(t),j𝒚(t)[in,jn]\displaystyle\leq 2\epsilon\frac{1}{T}\sum_{t=1}^{T}\bm{x}_{\mathsf{O}}^{(t)}\bm{y}_{\mathsf{O}}^{(t)}+(1+\gamma)\frac{1}{T}\sum_{t=1}^{T}\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i\leq n,j\leq n]
=1Tt=1T((1+γ)((1𝒙𝖮(t))(1𝒚𝖮(t))δ(t))+2ϵ𝒙𝖮(t)𝒚𝖮(t)),\displaystyle=\frac{1}{T}\sum_{t=1}^{T}\left((1+\gamma)((1-\bm{x}^{(t)}_{\mathsf{O}})(1-\bm{y}^{(t)}_{\mathsf{O}})-\delta^{(t)})+2\epsilon\bm{x}^{(t)}_{\mathsf{O}}\bm{y}^{(t)}_{\mathsf{O}}\right), (16)

where we have defined δ(t)i𝒙(t),j𝒚(t)[i2n,j2n]i𝒙(t),j𝒚(t)[in,jn]\delta^{(t)}\coloneqq\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i\leq 2n,j\leq 2n]-\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i\leq n,j\leq n] for each t[T]t\in[T]. Combining (15) and (16),

1Tt=1T(2ϵ(𝒙𝖮(t)+𝒚𝖮(t)2𝒙𝖮(t)𝒚𝖮(t))+(1+γ)(1𝒙𝖮(t)+𝒚𝖮(t)2(1𝒙𝖮(t))(1𝒚𝖮(t))+δ(t)))\displaystyle\frac{1}{T}\sum_{t=1}^{T}\left(2\epsilon\left(\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}-\bm{x}^{(t)}_{\mathsf{O}}\bm{y}^{(t)}_{\mathsf{O}}\right)+(1+\gamma)\left(1-\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}-(1-\bm{x}^{(t)}_{\mathsf{O}})(1-\bm{y}^{(t)}_{\mathsf{O}})+\delta^{(t)}\right)\right)
γ1Tt=1T(1𝒙𝖮(t)+𝒚𝖮(t)2).\displaystyle\leq\gamma\frac{1}{T}\sum_{t=1}^{T}\left(1-\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\right).

Thus, for ϵ<12\epsilon<\frac{1}{2},

1Tt=1T(𝒙𝖮(t)+𝒚𝖮(t)2𝒙𝖮(t)𝒚𝖮(t)+12δ(t))\displaystyle\frac{1}{T}\sum_{t=1}^{T}\left(\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}-\bm{x}^{(t)}_{\mathsf{O}}\bm{y}^{(t)}_{\mathsf{O}}+\frac{1}{2}\delta^{(t)}\right) γ2ϵ+1+γ1Tt=1T(1𝒙𝖮(t)+𝒚𝖮(t)2)\displaystyle\leq\frac{\gamma}{2\epsilon+1+\gamma}\frac{1}{T}\sum_{t=1}^{T}\left(1-\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\right)
γ1Tt=1T(1𝒙𝖮(t)+𝒚𝖮(t)2).\displaystyle\leq\gamma\frac{1}{T}\sum_{t=1}^{T}\left(1-\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\right).

Given that 𝒙𝖮(t),𝒚𝖮(t)[0,1]\bm{x}_{\mathsf{O}}^{(t)},\bm{y}^{(t)}_{\mathsf{O}}\in[0,1] and δ(t)0\delta^{(t)}\geq 0, it follows that for any t[T]t\in[T],

𝒙𝖮(t)+𝒚𝖮(t)2𝒙𝖮(t)𝒚𝖮(t)+12δ(t)γTδ(t)+𝒙𝖮(t)(1𝒚𝖮(t))+𝒚𝖮(t)(1𝒙𝖮(t))2γT.\frac{\bm{x}_{\mathsf{O}}^{(t)}+\bm{y}_{\mathsf{O}}^{(t)}}{2}-\bm{x}^{(t)}_{\mathsf{O}}\bm{y}^{(t)}_{\mathsf{O}}+\frac{1}{2}\delta^{(t)}\leq\gamma T\iff\delta^{(t)}+\bm{x}^{(t)}_{\mathsf{O}}(1-\bm{y}^{(t)}_{\mathsf{O}})+\bm{y}^{(t)}_{\mathsf{O}}(1-\bm{x}^{(t)}_{\mathsf{O}})\leq 2\gamma T.

That is, we have shown that for any t[T]t\in[T],

i𝒙(t),j𝒚(t)[(in,jn)(i=𝖮,j=𝖮)]12γT.\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[(i\leq n,j\leq n)\lor(i=\mathsf{O},j=\mathsf{O})]\geq 1-2\gamma T. (17)

Now, from the assumption that 𝝁\bm{\mu} is not supported only on (𝖮,𝖮)(\mathsf{O},\mathsf{O}), it follows that there exists t[T]t\in[T] such that 𝒙𝖮(t)1𝒚𝖮(t)1\bm{x}^{(t)}_{\mathsf{O}}\neq 1\lor\bm{y}^{(t)}_{\mathsf{O}}\neq 1. In particular, by (14), there exists such tt with the property that

2r(1𝒙𝖮(t)+𝒚𝖮(t)2)+2ϵ𝒙𝖮(t)+𝒚𝖮(t)22ϵ𝒙𝖮(t)𝒚𝖮(t)+i𝒙(t),j𝒚(t)[in,jn]𝖲𝖶(𝒙^(t)𝒚^(t)),2r\left(1-\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\right)+2\epsilon\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\leq 2\epsilon\bm{x}^{(t)}_{\mathsf{O}}\bm{y}^{(t)}_{\mathsf{O}}+\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i\leq n,j\leq n]\operatorname{\mathsf{SW}}(\widehat{\bm{x}}^{(t)}\otimes\widehat{\bm{y}}^{(t)}), (18)

where again we have defined Δn𝒙^in(t)=𝒙in(t)/i=1n𝒙i(t)\Delta^{n}\ni\widehat{\bm{x}}_{i\leq n}^{(t)}=\nicefrac{{\bm{x}^{(t)}_{i\leq n}}}{{\sum_{i=1}^{n}\bm{x}_{i}^{(t)}}} and Δn𝒚^jn(t)=𝒚jn(t)/j=1n𝒚j(t)\Delta^{n}\ni\widehat{\bm{y}}_{j\leq n}^{(t)}=\nicefrac{{\bm{y}^{(t)}_{j\leq n}}}{{\sum_{j=1}^{n}\bm{y}_{j}^{(t)}}}; any tt such that 𝒙𝖮(t)1𝒚𝖮(t)1\bm{x}^{(t)}_{\mathsf{O}}\neq 1\lor\bm{y}^{(t)}_{\mathsf{O}}\neq 1 that satisfies (18) must also satisfy i=1n𝒙i(t)>0\sum_{i=1}^{n}\bm{x}_{i}^{(t)}>0 and j=1n𝒚j(t)>0\sum_{j=1}^{n}\bm{y}_{j}^{(t)}>0. Given that (𝒙𝖮(t)+𝒚𝖮(t))/2𝒙𝖮(t)𝒚𝖮(t)(\bm{x}_{\mathsf{O}}^{(t)}+\bm{y}^{(t)}_{\mathsf{O}})/2\geq\bm{x}^{(t)}_{\mathsf{O}}\bm{y}^{(t)}_{\mathsf{O}}, we have

2r(1𝒙𝖮(t)+𝒚𝖮(t)2)i𝒙(t),j𝒚(t)[in,jn]𝖲𝖶(𝒙^(t)𝒚^(t)).2r\left(1-\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\right)\leq\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i\leq n,j\leq n]\operatorname{\mathsf{SW}}(\widehat{\bm{x}}^{(t)}\otimes\widehat{\bm{y}}^{(t)}).

Further,

i𝒙(t),j𝒚(t)[in,jn](1𝒙𝖮(t))(1𝒚𝖮(t))(1𝒙𝖮(t)+𝒚𝖮(t)2)0.\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i\leq n,j\leq n]\leq(1-\bm{x}_{\mathsf{O}}^{(t)})(1-\bm{y}_{\mathsf{O}}^{(t)})\leq\left(1-\frac{\bm{x}_{\mathsf{O}}^{(t)}+\bm{y}_{\mathsf{O}}^{(t)}}{2}\right)\neq 0.

We thus conclude that 𝖲𝖶(𝒙^(t)𝒚^(t))2r=1+γT/k\operatorname{\mathsf{SW}}(\widehat{\bm{x}}^{(t)}\otimes\widehat{\bm{y}}^{(t)})\geq 2r=1+\gamma T/k. We will now argue that 𝒙^i(t),𝒚^j(t)2/\widehat{\bm{x}}_{i}^{(t)},\widehat{\bm{y}}_{j}^{(t)}\leq 2/\ell for any i,j[n]i,j\in[n]. Continuing from (18), we bound

(1+γTk)(1𝒙𝖮(t)+𝒚𝖮(t)2)+2ϵ𝒙𝖮(t)+𝒚𝖮(t)22ϵ𝒙𝖮(t)𝒚𝖮(t)+(1+γ)(1𝒙𝖮(t))(1𝒚𝖮(t)).\left(1+\gamma\frac{T}{k}\right)\left(1-\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\right)+2\epsilon\frac{\bm{x}^{(t)}_{\mathsf{O}}+\bm{y}^{(t)}_{\mathsf{O}}}{2}\leq 2\epsilon\bm{x}^{(t)}_{\mathsf{O}}\bm{y}^{(t)}_{\mathsf{O}}+(1+\gamma)(1-\bm{x}^{(t)}_{\mathsf{O}})(1-\bm{y}^{(t)}_{\mathsf{O}}).

Thus,

(1+γ)𝒙𝖮(t)+𝒚𝖮(t)2𝒙𝖮(t)𝒚𝖮(t)γ,(1+\gamma)\frac{\bm{x}_{\mathsf{O}}^{(t)}+\bm{y}_{\mathsf{O}}^{(t)}}{2}-\bm{x}^{(t)}_{\mathsf{O}}\bm{y}^{(t)}_{\mathsf{O}}\leq\gamma,

By applying the AM-GM inequality, we have

(1+γ)𝒙𝖮(t)𝒚𝖮(t)𝒙𝖮(t)𝒚𝖮(t)γ(1𝒙𝖮(t)𝒚𝖮(t))(𝒙𝖮(t)𝒚𝖮(t)γ)0𝒙𝖮(t)𝒚𝖮(t)γ,(1+\gamma)\sqrt{\bm{x}_{\mathsf{O}}^{(t)}\bm{y}_{\mathsf{O}}^{(t)}}-\bm{x}_{\mathsf{O}}^{(t)}\bm{y}^{(t)}_{\mathsf{O}}\leq\gamma\iff\left(1-\sqrt{\bm{x}_{\mathsf{O}}^{(t)}\bm{y}_{\mathsf{O}}^{(t)}}\right)\left(\sqrt{\bm{x}_{\mathsf{O}}^{(t)}\bm{y}_{\mathsf{O}}^{(t)}}-\gamma\right)\leq 0\iff\sqrt{\bm{x}_{\mathsf{O}}^{(t)}\bm{y}_{\mathsf{O}}^{(t)}}\leq\gamma,

where we used the fact that 𝒙𝖮(t)1𝒚𝖮(t)1\bm{x}^{(t)}_{\mathsf{O}}\neq 1\lor\bm{y}^{(t)}_{\mathsf{O}}\neq 1. Combining with (17), it follows that i𝒙(t),j𝒚(t)[in,jn]1γ22γT\operatorname{\mathbb{P}}_{i\sim\bm{x}^{(t)},j\sim\bm{y}^{(t)}}[i\leq n,j\leq n]\geq 1-\gamma^{2}-2\gamma T. This implies that i=1n𝒙i(t)1γ22γT\sum_{i=1}^{n}\bm{x}_{i}^{(t)}\geq 1-\gamma^{2}-2\gamma T and j=1n𝒚j(t)1γ22γT\sum_{j=1}^{n}\bm{y}_{j}^{(t)}\geq 1-\gamma^{2}-2\gamma T. Again by (17), it follows that the expected utility of each player under 𝝁\bm{\mu} is at most (1+γ)/2+γkT(1+\gamma)/2+\gamma kT. Moreover, since 𝝁\bm{\mu} is assumed to be a CCE, by considering the deviation of Player yy to column n+in+i, we have that for any i[n]i\in[n],

1+γ+2γkT2k21Tt=1T𝒙i(t)r1Tt=1T𝒙𝖮(t)k2T𝒙i(t)T(1γ22γT)Tr𝒙^i(t)2Tk,\frac{1+\gamma+2\gamma kT}{2}\geq\frac{k}{2}\frac{1}{T}\sum_{t=1}^{T}\bm{x}_{i}^{(t)}-r\frac{1}{T}\sum_{t=1}^{T}\bm{x}^{(t)}_{\mathsf{O}}\geq\frac{k}{2T}\bm{x}_{i}^{(t)}-\frac{T-(1-\gamma^{2}-2\gamma T)}{T}r\implies\widehat{\bm{x}}_{i}^{(t)}\leq\frac{2T}{k},

when γ\gamma is sufficiently small. Finally, we let S{i[n]:𝒙^i(t),𝒚^i(t)1/(16)}S\coloneqq\{i\in[n]:\widehat{\bm{x}}^{(t)}_{i},\widehat{\bm{y}}^{(t)}_{i}\geq 1/(16\ell)\}, where k/T\ell\coloneqq k/T. As in the proof of Theorem 3.9, it follows that |S|/8|S|\geq\ell/8, and SS is a clique on GG. As a result, when GG does not contain a clique of size k/(8T)k/(8T), (𝖮,𝖮)(\mathsf{O},\mathsf{O}) is the only uniform TT-sparse CCE. This completes the proof. ∎

An interesting question, which is left for future work, is whether the construction of Theorem 3.11 can be extended to hold for non-uniform sparse CCE.

4 Hardness for low-precision equilibria

So far, we have established lower bounds for OptimalSparseCCE in the regime where both the equilibrium and the optimality gap are 𝗉𝗈𝗅𝗒(1/n)\mathsf{poly}(1/n). No-regret learning is often employed in the low-precision regime, where the precision scales as 1/\polylogn1/\polylog n. We will show here how to obtain hardness results in the latter regime, establishing Theorem 1.2 presented earlier in the introduction.

In that setting, we first remark that, even for 11-sparse CCE (that is, Nash equilibria), [LMM03] famously gave a quasipolynomial-time algorithm based on enumerating all possible strategies of support O(𝗅𝗈𝗀n/ϵ2)O(\mathsf{log}n/\epsilon^{2}). It is thus unlikely—under our current understanding of complexity—that the induced problem is \NP-hard. Instead, we will rely on 2.5 pertaining to the planted clique problem.

As before, our first lower bound (Theorem 4.2) makes use of Algorithm 1, but with GG now being the input of the planted clique problem (presented in Section 2.2). We will use the following lemma, which was observed in the form below by [HK11].

Lemma 4.1 ([HK11], \citeyearHazan11:How).

Suppose that there is randomized polynomial-time algorithm that given a planted clique problem with kc𝗅𝗈𝗀nk\geq c\mathsf{log}n, for a constant c>0c>0, finds a clique of size 100𝗅𝗈𝗀n100\mathsf{log}n with probability at least 1/21/2. Then, there is a randomized polynomial-time algorithm that solves the planted clique problem for any kc0𝗅𝗈𝗀nk\geq c_{0}\mathsf{log}n, for a constant c0>0c_{0}>0, with high probability.

Consider now any sparsity parameter T=\polylognT=\polylog n, and define k=100T𝗅𝗈𝗀nk=100T\mathsf{log}n. By the guarantee given in Section 3.1.1 concerning Algorithm 1, we know that we can then find a clique of size 100𝗅𝗈𝗀n100\mathsf{log}n by invoking an oracle to OptimalSparseCCE(𝒢,T,(𝗅𝗈𝗀n)c,(𝗅𝗈𝗀n)c)\textsc{OptimalSparseCCE}({\mathcal{G}},T,(\mathsf{log}n)^{-c},(\mathsf{log}n)^{-c}), where 𝒢=𝒢(G){\mathcal{G}}={\mathcal{G}}(G) is defined as in Algorithm 1 and cc is a sufficiently large constant (that depends on the sparsity). In turn, Lemma 4.1 above implies that we can solve the planted clique problem (w.h.p.). We thus arrive at the following.

Theorem 4.2.

Assuming that 2.5 holds, OptimalSparseCCE(𝒢,T,(𝗅𝗈𝗀n)c,(𝗅𝗈𝗀n)c)\textsc{OptimalSparseCCE}({\mathcal{G}},T,(\mathsf{log}n)^{-c},(\mathsf{log}n)^{-c}) requires time nΩ(𝗅𝗈𝗀n)n^{\Omega(\mathsf{log}n)} for any sparsity T=\polylognT=\polylog n and some constant c=c(T)c=c(T).

Finally, we show hardness in the regime where the equilibrium and the optimality gap are both constants using a different approach. Namely, we extend the reduction of [HK11], which was developed with Nash equilibria in mind. The main result is recalled below.

Theorem 4.3.

Assuming that 2.5 holds, OptimalSparseCCE(𝒢,T,c,c)\textsc{OptimalSparseCCE}({\mathcal{G}},T,c,c) requires time nΩ(𝗅𝗈𝗀n)n^{\Omega(\mathsf{log}n)} for any sparsity T=O(1)T=O(1) and some constant c=c(T)c=c(T).

We dedicate the rest of this section to the proof of Theorem 4.3.

4.1 Proof of Theorem 4.3

The proof mostly follows the argument of [HK11], but with certain modifications. Similarly to (3), we consider a two-player game defined by the payoff matrices

N×N𝐑12(𝐀𝐁𝐁𝟎(Nn)×(Nn)) and N×N𝐂12(𝐀𝐁𝐁𝟎(Nn)×(Nn)),{\mathbb{R}}^{N\times N}\ni\mathbf{R}\coloneqq\frac{1}{2}\begin{pmatrix}\mathbf{A}&-\mathbf{B}^{\top}\\ \mathbf{B}&\mathbf{0}_{(N-n)\times(N-n)}\end{pmatrix}\text{ and }{\mathbb{R}}^{N\times N}\ni\mathbf{C}\coloneqq\frac{1}{2}\begin{pmatrix}\mathbf{A}&\mathbf{B}^{\top}\\ -\mathbf{B}&\mathbf{0}_{(N-n)\times(N-n)}\end{pmatrix}, (19)

where NnN\gg n is sufficiently large (to be specified in the sequel), 𝐀\mathbf{A} is again the adjacency matrix (with 11 on the diagonal entries) of GG—the input of the planted clique problem, and 𝐁(Nn)×n\mathbf{B}\in{\mathbb{R}}^{(N-n)\times n} is defined by selecting each entry independently as

𝐁i,j={Mwith probability 341M, and 0otherwise.\mathbf{B}_{i,j}=\begin{cases}M&\text{with probability }\frac{3}{4}\frac{1}{M},\text{ and }\\ 0&\text{otherwise}.\end{cases}

Here, MM is a sufficiently large parameter that will be related to the sparsity (see Lemma 4.7). In what follows, we recall that kk\in{\mathbb{N}} represents the size of the planted clique. We first note that (𝐑,𝐂)(\mathbf{R},\mathbf{C}) in (19) admits, with high probability, a Nash equilibrium with welfare 11. This proof of completeness is similar to the one by [HK11].

Lemma 4.4.

Consider (𝐑,𝐂)(\mathbf{R},\mathbf{C}) with k96M2𝗅𝗈𝗀Nk\geq 96M^{2}\mathsf{log}N. Then, with high probability, there is a Nash equilibrium with welfare 11.

Proof.

We consider the strategy profile (𝒙,𝒚)(\bm{x},\bm{y}) in which both players play uniformly at random over the nodes corresponding to the planted clique. It is clear that 𝖲𝖶(𝒙𝒚)=1\operatorname{\mathsf{SW}}(\bm{x}\otimes\bm{y})=1. It thus suffices to show that (𝒙,𝒚)(\bm{x},\bm{y}) is indeed a Nash equilibrium of (19). By symmetry, we can analyze only deviations of Player xx. When ini\leq n, it is clear that there is no profitable deviation for Player xx as that player obtains 1/21/2 under (𝒙,𝒚)(\bm{x},\bm{y}), which is the highest utility attainable in the n×nn\times n submatrix corresponding to 𝐀\mathbf{A}. Let us consider thus some i>ni>n. The obtained utility then is half of the average over kk entries of 𝐁\mathbf{B}, which is highly concentrated around its the mean value. In particular, we will use the following Chernoff–Hoeffding bound [Hoe63]. Let X1,,XkX_{1},\dots,X_{k} be independent random variables such that |X1|,,|Xk|C|X_{1}|,\dots,|X_{k}|\leq C (almost surely). If X¯\bar{X} denotes their average, it holds that [X¯𝔼[X¯]t]ekt22C2\operatorname{\mathbb{P}}[\bar{X}-{\mathbb{E}}[\bar{X}]\geq t]\leq e^{-\frac{kt^{2}}{2C^{2}}} for any t>0t>0. In our case, if we set C=MC=M and t=1/4t=1/4, we find that the probability that ii is profitable is upper bounded by ek32M2e^{-\frac{k}{32M^{2}}}. Thus, by a union bound, the probability that there is a profitable deviation among i>ni>n is upper bounded by (Nn)ek32M21/n2(N-n)e^{-\frac{k}{32M^{2}}}\leq 1/n^{2}, where we used the assumption that k96M2𝗅𝗈𝗀Nk\geq 96M^{2}\mathsf{log}N. This completes the proof. ∎

We proceed by establishing soundness. The following lemma mirrors Lemma 3.5 that was shown earlier, and follows from the same observation.

Lemma 4.5.

Given a TT-sparse ϵ\epsilon-CCE of (𝐑,𝐂)(\mathbf{R},\mathbf{C}) with welfare 1ϵ^1-\hat{\epsilon}, we can compute in polynomial time an (ϵ+2Mϵ^)(\epsilon+2M\hat{\epsilon})-CCE of (𝐑,𝐂)(\mathbf{R},\mathbf{C}) with welfare at least 1ϵ^1-\hat{\epsilon} and supported only on [n]×[n][n]\times[n].

The proof of this lemma is similar to that of Lemma 3.5. In particular, if we define δ(t)1i=1n𝒙i(t)j=1n𝒚j(t)\delta^{(t)}\coloneqq 1-\sum_{i=1}^{n}\bm{x}_{i}^{(t)}\sum_{j=1}^{n}\bm{y}_{j}^{(t)} for each t[T]t\in[T], we have

1ϵ^t=1Tα(t)𝖲𝖶(𝒙(t)𝒚(t))=t=1Tα(t)i=1nj=1n𝒙i(t)𝐀i,j𝒚j(t)1t=1Tα(t)δ(t),1-\hat{\epsilon}\leq\sum_{t=1}^{T}\alpha^{(t)}\operatorname{\mathsf{SW}}(\bm{x}^{(t)}\otimes\bm{y}^{(t)})=\sum_{t=1}^{T}\alpha^{(t)}\sum_{i=1}^{n}\sum_{j=1}^{n}\bm{x}_{i}^{(t)}\mathbf{A}_{i,j}\bm{y}_{j}^{(t)}\leq 1-\sum_{t=1}^{T}\alpha^{(t)}\delta^{(t)},

in turn implying that t=1Tα(t)δ(t)ϵ^\sum_{t=1}^{T}\alpha^{(t)}\delta^{(t)}\leq\hat{\epsilon}. The rest of the argument then is the same as in Lemma 3.5, giving the conclusion of Lemma 4.5.

Next, the lemma below gives a bound analogous to Lemma 3.6. (In what follows, we can safely assume that c2𝗅𝗈𝗀nc_{2}\mathsf{log}n is an integer.)

Lemma 4.6.

Let 𝛍=t=1Tα(t)(𝐱(t)𝐲(t))\bm{\mu}=\sum_{t=1}^{T}\alpha^{(t)}(\bm{x}^{(t)}\otimes\bm{y}^{(t)}) be a 12\frac{1}{2}-CCE of (𝐑,𝐂)(\mathbf{R},\mathbf{C}) such that each 𝐱(t)𝐲(t)\bm{x}^{(t)}\otimes\bm{y}^{(t)} is supported only on [n]×[n][n]\times[n]. Suppose further that N=nc12nN=n^{c_{1}}\geq 2n and c1c2𝗅𝗈𝗀(4M/3)2c_{1}-c_{2}\mathsf{log}(4M/3)\geq 2. Then, with high probability, the overall probability mass every player places on every subset of c2𝗅𝗈𝗀nc_{2}\mathsf{log}n pure strategies of the ttth product distribution is at most 2/(Mα(t))2/(M\alpha^{(t)}).

Proof.

Let dc2𝗅𝗈𝗀nd\coloneqq c_{2}\mathsf{log}n. For the sake of contradiction, suppose that Player yy places on some set of c2𝗅𝗈𝗀nc_{2}\mathsf{log}n pure strategies of the ttth product distribution more than 2/(Mα(t))2/(M\alpha^{(t)}) probability mass. We will compute the probability that there is a row of 𝐁\mathbf{B} in which all entries corresponding to dd have a value of MM. If that event happens, Player xx can deviate to that row and obtain a utility larger than 11, while the current utility is at most 1/21/2. The probability that event occurs for a single row is pdp^{d} for p34Mp\coloneqq\frac{3}{4M}, but 𝐁\mathbf{B} has NnN-n rows each with independent randomization (by construction of 𝐁\mathbf{B}). Thus, the probability that no such rows exists can be bounded as

(1pd)Nnepd(Nn)=e(Nn)nc2𝗅𝗈𝗀(4M/3)en22,(1-p^{d})^{N-n}\leq e^{-p^{d}(N-n)}=e^{-(N-n)n^{-c_{2}\mathsf{log}(4M/3)}}\leq e^{-\frac{n^{2}}{2}},

where we used the assumption that N=nc12nN=n^{c_{1}}\geq 2n and c1c2𝗅𝗈𝗀(4M/3)2c_{1}-c_{2}\mathsf{log}(4M/3)\geq 2. Finally, we can rule out all subsets with size c2𝗅𝗈𝗀nc_{2}\mathsf{log}n via a union bound, leading to the claim. ∎

In the lemma below, the density of sets S,T[n]S,T\subseteq[n] with |S|=|T||S|=|T|, denoted by dens(S,T)\mathrm{dens}(S,T), is the probability that an edge exists (with self-loops included) between two nodes chosen uniformly at random from SS and TT.

Lemma 4.7.

Let M>4TM>4T and ϵ^1/(10T)\hat{\epsilon}\leq 1/(10T). Given a 12\frac{1}{2}-CCE of (𝐑,𝐂)(\mathbf{R},\mathbf{C}) supported only on [n]×[n][n]\times[n] with welfare at least 1ϵ^1-\hat{\epsilon}, we can find in polynomial time sets S,T[n]S,T\subseteq[n] with |S|=|T|=c2𝗅𝗈𝗀n|S|=|T|=c_{2}\mathsf{log}n and dens(S,T)35\mathrm{dens}(S,T)\geq\frac{3}{5}.

Proof.

Let 𝝁=t=1Tα(t)(𝒙(t)𝒚(t))\bm{\mu}=\sum_{t=1}^{T}\alpha^{(t)}(\bm{x}^{(t)}\otimes\bm{y}^{(t)}), where with a slight abuse of notation we let 𝒙(t),𝒚(t)Δn\bm{x}^{(t)},\bm{y}^{(t)}\in\Delta^{n} for all t[T]t\in[T] (since 𝝁\bm{\mu} is assumed to be supported only on [n]×[n][n]\times[n]). We select any t[T]t\in[T] such that α(t)1/T\alpha^{(t)}\geq 1/T. Since it is assumed that 𝝁\bm{\mu} has welfare at least 1ϵ^1-\hat{\epsilon}, it follows that

𝒙(t),𝐀𝒚(t)1α(t)(𝖲𝖶(𝝁)τtα(τ)𝖲𝖶(𝒙(τ)𝒚(τ)))1δ,\langle\bm{x}^{(t)},\mathbf{A}\bm{y}^{(t)}\rangle\geq\frac{1}{\alpha^{(t)}}\left(\operatorname{\mathsf{SW}}(\bm{\mu})-\sum_{\tau\neq t}\alpha^{(\tau)}\operatorname{\mathsf{SW}}(\bm{x}^{(\tau)}\otimes\bm{y}^{(\tau)})\right)\geq 1-\delta,

where δ=ϵ^T\delta=\hat{\epsilon}T. Let Iy(t){j:supp(𝒚(t)):𝒙(t),𝐀𝒆j45}I_{y}^{(t)}\coloneqq\{j:\text{supp}(\bm{y}^{(t)}):\langle\bm{x}^{(t)},\mathbf{A}\bm{e}_{j}\rangle\geq\frac{4}{5}\}, where 𝒆j\bm{e}_{j} is the jjth unit vector. Then, jIy(t)𝒚j(t)15δ\sum_{j\in I_{y}^{(t)}}\bm{y}_{j}^{(t)}\geq 1-5\delta, for otherwise 𝒙(t),𝐀𝒚(t)<(15δ)+(5δ)45=1δ\langle\bm{x}^{(t)},\mathbf{A}\bm{y}^{(t)}\rangle<(1-5\delta)+(5\delta)\frac{4}{5}=1-\delta. Using the fact that ϵ^110T<15T25M\hat{\epsilon}\leq\frac{1}{10T}<\frac{1}{5T}-\frac{2}{5M} (since M>4TM>4T), it follows that jIy(t)𝒚j(t)>2TM2Mα(t)\sum_{j\in I_{y}^{(t)}}\bm{y}_{j}^{(t)}>\frac{2T}{M}\geq\frac{2}{M\alpha^{(t)}}. As a result, Lemma 4.6 tells us that |Iy(t)|c2𝗅𝗈𝗀n|I_{y}^{(t)}|\geq c_{2}\mathsf{log}n. If |Iy(t)|>c2𝗅𝗈𝗀n|I_{y}^{(t)}|>c_{2}\mathsf{log}n, we let Iy(t)I_{y}^{(t)} be an arbitrary subset of Iy(t)I_{y}^{(t)} with size c2𝗅𝗈𝗀nc_{2}\mathsf{log}n. We now consider 𝒖(Iy(t))Δn\bm{u}(I_{y}^{(t)})\in\Delta^{n}, the uniform distribution over Iy(t)I_{y}^{(t)}. We have

𝒙(t),𝐀𝒖(Iy(t))45.\langle\bm{x}^{(t)},\mathbf{A}\bm{u}(I_{y}^{(t)})\rangle\geq\frac{4}{5}.

Similarly, we define Ix(t){isupp(𝒙(t)):𝒆i,𝐀𝒖(Iy(t))35}I_{x}^{(t)}\coloneqq\{i\in\text{supp}(\bm{x}^{(t)}):\langle\bm{e}_{i},\mathbf{A}\bm{u}(I_{y}^{(t)})\rangle\geq\frac{3}{5}\}. Then, iIx(t)𝒙i(t)12>2TM\sum_{i\in I_{x}^{(t)}}\bm{x}_{i}^{(t)}\geq\frac{1}{2}>\frac{2T}{M} (since M>4TM>4T). Thus, Lemma 3.4 implies that |Ix(t)|c2𝗅𝗈𝗀n|I_{x}^{(t)}|\geq c_{2}\mathsf{log}n. If |Ix(t)|>c2𝗅𝗈𝗀n|I_{x}^{(t)}|>c_{2}\mathsf{log}n, we let Ix(t)I_{x}^{(t)} be an arbitrary subset of Ix(t)I_{x}^{(t)} with size c2𝗅𝗈𝗀nc_{2}\mathsf{log}n. Thus, we have found sets Ix(t),Iy(t)[n]I_{x}^{(t)},I_{y}^{(t)}\subseteq[n] with |Ix(t)|=|Iy(t)|=c2𝗅𝗈𝗀n|I_{x}^{(t)}|=|I_{y}^{(t)}|=c_{2}\mathsf{log}n and

𝒖(Ix(t)),𝐀𝒖(Iy(t))35;\langle\bm{u}(I_{x}^{(t)}),\mathbf{A}\bm{u}(I_{y}^{(t)})\rangle\geq\frac{3}{5};

that is, dens(Ix(t),Iy(t))35\mathrm{dens}(I_{x}^{(t)},I_{y}^{(t)})\geq\frac{3}{5}. ∎

We finally conclude the proof of soundness by stating a lemma shown by [HK11], which we include without proof.

Lemma 4.8 ([HK11], \citeyearHazan11:How).

Given sets S,T[n]S,T\subseteq[n] with |S|=|T|=2000𝗅𝗈𝗀n|S|=|T|=2000\mathsf{log}n and dens(S,T)35\mathrm{dens}(S,T)\geq\frac{3}{5}, we can find in polynomial time a clique of size 100𝗅𝗈𝗀n100\mathsf{log}n with high probability.

As a result, Theorem 4.3 follows from Lemmas 4.8, 4.7, 4.6, 4.5 and 4.1 by making the following choice of parameters:

  • M5TM\coloneqq 5T;

  • c22000c_{2}\coloneqq 2000;

  • c1c2𝗅𝗈𝗀(4M/3)+2c_{1}\coloneqq\lceil c_{2}\mathsf{log}(4M/3)\rceil+2 and Nnc1N\coloneqq n^{c_{1}};

  • k96M2𝗅𝗈𝗀Nk\coloneqq\lceil 96M^{2}\mathsf{log}N\rceil;

  • ϵ^ϵ/(4M)\hat{\epsilon}\coloneqq\epsilon/(4M); and

  • ϵ1/4\epsilon\coloneqq 1/4.

This reduction can also be used when T=ω(1)T=\omega(1), but then the required precision for ϵ\epsilon and ϵ^\hat{\epsilon} would need to be accordingly small (we recall that one has to ultimately divide the entries of (19) by M=Θ(T)M=\Theta(T) so as to normalize the utilities in [1,1][-1,1], which would result in ϵ=Θ(1/T)\epsilon=\Theta(1/T)). It is not clear how to overcome this issue and prove stronger lower bounds with the current approach, but we suspect that such lower bounds should apply.

5 Future research

Our results raise a number of interesting avenues for future research. First, does Theorem 3.11 apply with respect to non-uniform sparse CCE? That would readily generalize Corollaries 3.13, 3.12 and 3.14. Turning to the low-precision regime, what is the precise threshold of computational tractability for OptimalSparseCCE? Theorem 4.3 only precludes T=O(1)T=O(1) (under constant precision), but we suspect that it can be significantly strengthened. Furthermore, it seems likely that one can instead prove lower bounds by relying on the exponential-time hypothesis (ETH)—as opposed to the planted clique conjecture (cf. [BKW15]), which would provide further evidence for the intractability of the problem. Yet, extending existing ETH-based reductions to sparse CCE appears to introduce considerable challenges. Last but not least, perhaps the most important question concerns the complexity of computing sparse CCE without imposing further constraints, a problem that will likely require quite different techniques from the ones employed here.

Acknowledgments

This material is based on work supported by the Vannevar Bush Faculty Fellowship ONR N00014-23-1-2876, National Science Foundation grants RI-2312342 and RI-1901403, ARO award W911NF2210266, and NIH award A240108S001. Alkis Kalavasis was supported by the Institute for Foundations of Data Science at Yale.

References

  • [ABC13] Per Austrin, Mark Braverman and Eden Chlamtac “Inapproximability of NP-Complete Variants of Nash Equilibrium” In Theory Comput. 9, 2013, pp. 117–142
  • [Abe+18] Jacob Abernethy, Kevin A Lai, Kfir Y Levy and Jun-Kun Wang “Faster rates for convex-concave games” In Conference on Learning Theory (COLT), 2018
  • [ABW10] Benny Applebaum, Boaz Barak and Avi Wigderson “Public-key cryptography from different assumptions” In Proceedings of the Annual Symposium on Theory of Computing (STOC), 2010
  • [AKS98] Noga Alon, Michael Krivelevich and Benny Sudakov “Finding a large hidden clique in a random graph” In Random Structures & Algorithms 13.3-4, 1998, pp. 457–466
  • [Alo+07] Noga Alon et al. “Testing k-wise and almost k-wise independence” In Proceedings of the Annual Symposium on Theory of Computing (STOC), 2007
  • [Aro+98] Sanjeev Arora et al. “Proof Verification and the Hardness of Approximation Problems” In Journal of the ACM 45.3, 1998, pp. 501–555
  • [Aum74] Robert Aumann “Subjectivity and Correlation in Randomized Strategies” In Journal of Mathematical Economics 1, 1974, pp. 67–96
  • [Bak+22] Anton Bakhtin et al. “Human-level play in the game of <i>Diplomacy</i> by combining language models with strategic reasoning” In Science 378.6624, 2022, pp. 1067–1074
  • [Bal+13] Maria-Florina Balcan et al. “Finding endogenously formed communities” In Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), 2013
  • [Bal+14] Maria-Florina Balcan, Sara Krehbiel, Georgios Piliouras and Jinwoo Shin “Near-Optimality in Covering Games by Exposing Global Information” In ACM Transaction on Economics and Computation (TEAC) 2.4, 2014, pp. 13:1–13:22
  • [Bar+19] Boaz Barak et al. “A nearly tight sum-of-squares lower bound for the planted clique problem” In SIAM Journal on Computing 48.2, 2019, pp. 687–735
  • [BB15] Yakov Babichenko and Siddharth Barman “Query Complexity of Correlated Equilibrium” In ACM Trans. Economics and Comput. 3.4, 2015, pp. 22:1–22:9
  • [BBM13] Maria-Florina Balcan, Avrim Blum and Yishay Mansour “Circumventing the Price of Anarchy: Leading Dynamics to Good Behavior” In SIAM Journal on Computing 42.1, 2013, pp. 230–264
  • [BBP14] Yakov Babichenko, Siddharth Barman and Ron Peretz “Simple approximate equilibria in large games” In Proceedings of the ACM Conference on Economics and Computation (EC), 2014
  • [BGS98] Mihir Bellare, Oded Goldreich and Madhu Sudan “Free Bits, PCPs, and Nonapproximability-Towards Tight Results” In SIAM Journal on Computing, 1998
  • [BKW15] Mark Braverman, Young Kun-Ko and Omri Weinstein “Approximating the best Nash Equilibrium in no{}^{\mbox{o}}(log n){}^{\mbox{(log \emph{n})}}-time breaks the Exponential Time Hypothesis” In Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), 2015
  • [BL15] Siddharth Barman and Katrina Ligett “Finding any nontrivial coarse correlated equilibrium is hard” In ACM SIGecom Exchanges 14.1, 2015, pp. 76–79
  • [Bla56] David Blackwell “An analog of the minmax theorem for vector payoffs” In Pacific Journal of Mathematics 6, 1956, pp. 1–8
  • [Blu+08] Avrim Blum, MohammadTaghi Hajiaghayi, Katrina Ligett and Aaron Roth “Regret minimization and the price of total anarchy” In Proceedings of the Annual Symposium on Theory of Computing (STOC), 2008
  • [Bow+15] Michael Bowling, Neil Burch, Michael Johanson and Oskari Tammelin “Heads-up Limit Hold’em Poker is Solved” In Science 347.6218, 2015, pp. 145–149
  • [BR13] Quentin Berthet and Philippe Rigollet “Complexity Theoretic Lower Bounds for Sparse Principal Component Detection” In Conference on Learning Theory (COLT), 2013
  • [BS19] Noam Brown and Tuomas Sandholm “Solving imperfect-information games via discounted regret minimization” In Conference on Artificial Intelligence (AAAI), 2019
  • [BS19a] Noam Brown and Tuomas Sandholm “Superhuman AI for multiplayer poker” In Science 365.6456, 2019, pp. 885–890
  • [BS94] Mihir Bellare and Madhu Sudan “Improved non-approximability results” In Proceedings of the Annual Symposium on Theory of Computing (STOC), 1994
  • [Cev+23] Volkan Cevher et al. “Alternation makes the adversary weaker in two-player games” In Proceedings of the Annual Conference on Neural Information Processing Systems (NeurIPS), 2023
  • [CL06] Nicolo Cesa-Bianchi and Gábor Lugosi “Prediction, learning, and games” Cambridge university press, 2006
  • [CS08] Vincent Conitzer and Tuomas Sandholm “New Complexity Results about Nash Equilibria” Early version in IJCAI-03 In Games and Economic Behavior 63.2, 2008, pp. 621–641
  • [DDK15] Constantinos Daskalakis, Alan Deckelbaum and Anthony Kim “Near-optimal no-regret algorithms for zero-sum games” In Games and Economic Behavior 92 Elsevier, 2015, pp. 327–348
  • [DFG21] Constantinos Daskalakis, Maxwell Fishelson and Noah Golowich “Near-Optimal No-Regret Learning in General Games” In Proceedings of the Annual Conference on Neural Information Processing Systems (NeurIPS), 2021
  • [Fei+96] Uriel Feige et al. “Interactive Proofs and the Hardness of Approximating Cliques” In J. ACM 43.2, 1996, pp. 268–292
  • [FGK23] Dylan J. Foster, Noah Golowich and Sham M. Kakade “Hardness of Independent Learning and Sparse Equilibrium Computation in Markov Games” In International Conference on Machine Learning (ICML), 2023
  • [FNS07] Tomás Feder, Hamid Nazerzadeh and Amin Saberi “Approximating nash equilibria using small-support strategies” In Proceedings of the ACM Conference on Economics and Computation (EC), 2007
  • [FV97] Dean Foster and Rakesh Vohra “Calibrated Learning and Correlated Equilibrium” In Games and Economic Behavior 21, 1997, pp. 40–55
  • [GK23] Paul W. Goldberg and Matthew J. Katzman “Lower bounds for the query complexity of equilibria in Lipschitz games”, 2023, pp. 113931
  • [GR16] Paul W. Goldberg and Aaron Roth “Bounds for the Query Complexity of Approximate Equilibria” In ACM Transaction on Economics and Computation (TEAC) 4.4, 2016, pp. 24:1–24:25
  • [GZ89] Itzhak Gilboa and Eitan Zemel “Nash and Correlated Equilibria: Some Complexity Considerations” In Games and Economic Behavior 1, 1989, pp. 80–93
  • [Hås99] Johan Håstad “Clique is hard to approximate within n1ϵn^{1-\epsilon} In Acta Mathematica 182, 1999, pp. 105–142
  • [HK11] Elad Hazan and Robert Krauthgamer “How Hard Is It to Approximate the Best Nash Equilibrium?” In SIAM Journal on Computing 40.1, 2011, pp. 79–91
  • [HM00] Sergiu Hart and Andreu Mas-Colell “A Simple Adaptive Procedure Leading to Correlated Equilibrium” In Econometrica 68, 2000, pp. 1127–1150
  • [Hoe63] Wassily Hoeffding “Probability Inequalities for Sums of Bounded Random Variables” In Journal of the American Statistical Association 58.301, 1963, pp. 13–30
  • [Jer92] Mark Jerrum “Large cliques elude the Metropolis process” In Random Structures & Algorithms 3.4, 1992, pp. 347–359
  • [JPY88] David S. Johnson, Christos H. Papadimitriou and Mihalis Yannakakis “How Easy is Local Search?” In J. Comput. Syst. Sci. 37.1, 1988, pp. 79–100
  • [Kar72] Richard Karp “Reducibility among Combinatorial Problems” In Complexity of Computer Computations NY: Plenum Press, 1972, pp. 85–103
  • [Kho01] S. Khot “Improved inapproximability results for MaxClique, chromatic number and approximate graph coloring” In Proceedings of the Annual Symposium on Foundations of Computer Science (FOCS), 2001
  • [KM18] Pravesh K. Kothari and Ruta Mehta “Sum-of-squares meets nash: lower bounds for finding any equilibrium” In Proceedings of the Annual Symposium on Theory of Computing (STOC), 2018
  • [KN22] Yoav Kolumbus and Noam Nisan “Auctions between Regret-Minimizing Agents” In The ACM Web Conference (WWW), 2022
  • [Kuč95] Luděk Kučera “Expected complexity of graph partitioning problems” In Discrete Applied Mathematics 57.2, 1995, pp. 193–212
  • [KYV20] David Kempe, Sixie Yu and Yevgeniy Vorobeychik “Inducing Equilibria in Networked Public Goods Games through Network Structure Modification” In Autonomous Agents and Multi-Agent Systems, 2020
  • [Li+20] Jiayang Li, Jing Yu, Yu Marco Nie and Zhaoran Wang “End-to-End Learning and Intervention in Games” In Proceedings of the Annual Conference on Neural Information Processing Systems (NeurIPS), 2020
  • [Liu+22] Boyi Liu et al. “Inducing Equilibria via Incentives: Simultaneous Design-and-Play Ensures Global Convergence” In Proceedings of the Annual Conference on Neural Information Processing Systems (NeurIPS), 2022
  • [LMM03] Richard Lipton, Evangelos Markakis and Aranyak Mehta “Playing Large Games Using Simple Strategies” In Proceedings of the ACM Conference on Electronic Commerce (ACM-EC), 2003
  • [LW94] Nick Littlestone and M.. Warmuth “The Weighted Majority Algorithm” In Information and Computation 108.2, 1994, pp. 212–261
  • [Mai+23] Arnab Maiti, Ross Boczar, Kevin G. Jamieson and Lillian J. Ratliff “Query-Efficient Algorithms to Find the Unique Nash Equilibrium in a Two-Player Zero-Sum Matrix Game” In CoRR abs/2310.16236, 2023
  • [Mgu+19] David Mguni et al. “Coordinating the Crowd: Inducing Desirable Equilibria in Non-Cooperative Systems” In Autonomous Agents and Multi-Agent Systems, 2019
  • [MV09] Lorenz Minder and Dan Vilenchik “Small clique detection and approximate Nash equilibria” In International Workshop on Approximation Algorithms for Combinatorial Optimization, 2009
  • [NST15] Denis Nekipelov, Vasilis Syrgkanis and Éva Tardos “Econometrics for Learning Agents” In Proceedings of the ACM Conference on Economics and Computation (EC), 2015
  • [Pap94] Christos H. Papadimitriou “On the complexity of the parity argument and other inefficient proofs of existence” In Journal of Computer and system Sciences 48.3 Elsevier, 1994, pp. 498–532
  • [Per+22] Julien Perolat et al. “Mastering the game of Stratego with model-free multiagent reinforcement learning” In Science 378.6623, 2022, pp. 990–996
  • [PR08] Christos H. Papadimitriou and Tim Roughgarden “Computing correlated equilibria in multi-player games” In Journal of the ACM 55.3, 2008, pp. 14:1–14:29
  • [PR24] Binghui Peng and Aviad Rubinstein “The complexity of approximate (coarse) correlated equilibrium for incomplete information games” In Conference on Learning Theory (COLT), 2024
  • [PSS22] Georgios Piliouras, Ryann Sim and Stratis Skoulakis “Beyond Time-Average Convergence: Near-Optimal Uncoupled Online Learning via Clairvoyant Multiplicative Weights Update” In Proceedings of the Annual Conference on Neural Information Processing Systems (NeurIPS), 2022
  • [Rob51] Julia Robinson “An iterative method of solving a game” In Annals of Mathematics 54, 1951, pp. 296–301
  • [Rou15] Tim Roughgarden “Intrinsic Robustness of the Price of Anarchy” In Journal of the ACM 62.5, 2015, pp. 32:1–32:42
  • [RS13] Alexander Rakhlin and Karthik Sridharan “Optimization, learning, and games with predictable sequences” In Proceedings of the Annual Conference on Neural Information Processing Systems (NeurIPS), 2013
  • [Syr+15] Vasilis Syrgkanis, Alekh Agarwal, Haipeng Luo and Robert E Schapire “Fast convergence of regularized learning in games” In Proceedings of the Annual Conference on Neural Information Processing Systems (NeurIPS), 2015
  • [Tam+15] Oskari Tammelin, Neil Burch, Michael Johanson and Michael Bowling “Solving Heads-up Limit Texas Hold’em” In Proceedings of the 24th International Joint Conference on Artificial Intelligence (IJCAI), 2015
  • [WTP22] Andre Wibisono, Molei Tao and Georgios Piliouras “Alternating Mirror Descent for Constrained Min-Max Games” In Proceedings of the Annual Conference on Neural Information Processing Systems (NeurIPS), 2022
  • [Zha+22] Brian Hu Zhang, Gabriele Farina, Andrea Celli and Tuomas Sandholm “Optimal Correlated Equilibria in General-Sum Extensive-Form Games: Fixed-Parameter Algorithms, Hardness, and Two-Sided Column-Generation” In Proceedings of the ACM Conference on Economics and Computation (EC), 2022
  • [ZLB22] Hugh Zhang, Adam Lerer and Noam Brown “Equilibrium Finding in Normal-Form Games via Greedy Regret Minimization” In Conference on Artificial Intelligence (AAAI), 2022
  • [Zuc07] David Zuckerman “Linear Degree Extractors and the Inapproximability of Max Clique and Chromatic Number” In Theory Comput. 3.1, 2007, pp. 103–128