This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Black-Box Strategies and Equilibrium for Games with Cumulative Prospect Theoretic Players

Soham R. Phade and Venkat Anantharam Research supported by the NSF grants CNS–1527846, CCF–1618145 and CCF-1901004, by the NSF Science & Technology Center for Science of Information Grant number CCF-0939370 and by the William and Flora Hewlett Foundation supported Center for Long Term Cybersecurity at Berkeley. The authors are with the Department of Electrical Engineering and Computer Sciences, University of California, Berkeley, Berkeley, CA 94720. soham_phade@berkeley.edu, ananth@eecs.berkeley.edu
Abstract

The betweenness property of preference relations states that a probability mixture of two lotteries should lie between them in preference. It is a weakened form of the independence property and hence satisfied in expected utility theory (EUT). Experimental violations of betweenness are well-documented and several preference theories, notably cumulative prospect theory (CPT), do not satisfy betweenness. We prove that CPT preferences satisfy betweenness if and only if they conform with EUT preferences. In game theory, lack of betweenness in the players’ preference relations makes it essential to distinguish between the two interpretations of a mixed action by a player – conscious randomizations by the player and the uncertainty in the beliefs of the opponents. We elaborate on this distinction and study its implication for the definition of Nash equilibrium. This results in four different notions of equilibrium, with pure and mixed action Nash equilibrium being two of them. We dub the other two pure and mixed black-box strategy Nash equilibrium respectively. We resolve the issue of existence of such equilibria and examine how these different notions of equilibrium compare with each other.

1 Introduction

There is a large amount of evidence that human agents as decision-makers do not conform to the independence axiom of expected utility theory (EUT). (See, for example, Allais (1953); Weber and Camerer (1987) and Machina (1992).) This has led to the study of several alternate theories that do away with the independence axiom (Machina, 2014). Amongst these, the cumulative prospect theory (CPT) of Tversky and Kahneman (1992) stands out since it accommodates many of the empirically observed behavioral features from human experiments without losing much analytical tractability (Wakker, 2010). Further, it includes EUT as a special case.

The independence axiom says that if lottery L1L_{1} is weakly preferred over lottery L2L_{2} by an agent (i.e. the agent wants lottery L1L_{1} at least as much as lottery L2L_{2}), and LL is some other lottery, then, for 0α10\leq\alpha\leq 1, the combined lottery αL1+(1α)L\alpha L_{1}+(1-\alpha)L is weakly preferred over the combined lottery αL2+(1α)L\alpha L_{2}+(1-\alpha)L by that agent. A weakened form of the independence axiom, called betweenness, says that if lottery L1L_{1} is weakly preferred over lottery L2L_{2} (by an agent), then, for any 0α10\leq\alpha\leq 1, the mixed lottery L=αL1+(1α)L2L=\alpha L_{1}+(1-\alpha)L_{2} must lie between the lotteries L1L_{1} and L2L_{2} in preference. Betweenness implies that if an agent is indifferent between L1L_{1} and L2L_{2}, then she is indifferent between any mixtures of them too. It is known that independence implies betweenness, but betweenness does not imply independence (Chew, 1989). As a result, EUT preferences, which are known to satisfy the independence axiom, also satisfy betweenness. CPT preferences, on the other hand, do not satisfy betweenness in general (see example 2.2). In fact, in theorem 2.3, we show that CPT preferences satisfy betweenness if and only if they are EUT preferences (recall that EUT preferences are a special case of CPT preferences). Several empirical studies show systematic violations of betweenness (Camerer and Ho, 1994; Agranov and Ortoleva, 2017; Dwenger et al., 2012; Sopher and Narramore, 2000), and this makes the use of CPT more attractive than EUT for modeling human preferences. Further evidence comes from Camerer and Ho (1994), where the authors fit data from nine studies using three non-EUT models, one of them being CPT, to find that, compared to the EUT model, the non-EUT models perform better.

Suppose in a non-cooperative game that given her beliefs about the other players, a player is indifferent between two of her actions. Then according to EUT, she should be indifferent between any of the mixtures of these two actions. This facilitates the proof of the existence of a Nash equilibrium in mixed actions for such games. However, with CPT preferences, the player could either prefer some mixture of these two actions over the individual actions or vice versa.

As a result, it is important to make a distinction in CPT regarding whether the players can actively randomize over their actions or not. One way to enable active randomization is by assuming that each player has access to a randomizing device and the player can “commit” to the outcome of this randomization. The commitment assumption is necessary, as is evident from the following scenario (the gambles presented below appear in Prelec (1990)). Alice needs to choose between the following two actions:

  1. 1.

    Action 11 results in a lottery L1={(0.34,$20,000);(0.66,$0)}L_{1}=\{(0.34,\$20{,}000);(0.66,\$0)\}, i.e. she receives $20,000\$20{,}000 with probability 0.340.34 and nothing with probability 0.660.66.

  2. 2.

    Action 22 results in a lottery L2={(0.17,$30,000);(0.83,$0)}L_{2}=\{(0.17,\$30{,}000);(0.83,\$0)\}.

(See example 3.7 for an instance of a 22-player game with Alice and Bob, where Alice has two actions that result in the above two lotteries.) Note that L1L_{1} is a less risky gamble with a lower reward and L2L_{2} is a more risky gamble with a higher reward. Now consider a compound lottery L=16/17L1+1/17L2L=16/17L_{1}+1/17L_{2}. Substituting for the lotteries L1L_{1} and L2L_{2} we get LL in its reduced form to be

L={(0.01,$30,000);(0.32,$20,000);(0.67,$0)}.L=\{(0.01,\$30{,}000);(0.32,\$20{,}000);(0.67,\$0)\}.

In example 2.1, we provide a CPT model for Alice’s preferences that result in lottery L1L_{1} being preferred over lottery L2L_{2}, whereas lottery LL is preferred over lotteries L1L_{1} and L2L_{2}. Roughly speaking, the underlying intuition is that Alice is risk-averse in general, and she prefers lottery L1L_{1} over lottery L2L_{2}. However, she overweights the small 1%1\% chance of getting $30,000\$30{,}000 in LL and finds it lucrative enough to make her prefer lottery LL over both the lotteries L1L_{1} and L2L_{2}. Let us say Alice has a biased coin that she can use to implement the randomized strategy. Now, if Alice tossed the coin, and the outcome was to play action 2, then in the absence of commitment, she will switch to action 1, since she prefers lottery L1L_{1} over lottery L2L_{2}. Commitment can be achieved, for example, by asking a trusted party to implement the randomized strategy for her or use a device that would carry out the randomization and implement the outcome without further consultation with Alice. Regardless of the implementation mechanism, we will call such randomized strategies black-box strategies. The above problem of commitment is closely related to the problem of using non-EUT models in dynamic decisions. For an interesting discussion on this topic, see Wakker (2010, Appendix C) and the references therein.

Traditionally, mixed actions have been considered from two viewpoints, especially in the context of mixed action Nash equilibrium. According to the first viewpoint, these are conscious randomizations by the players – each player only knows her mixed action and not its pure realization. The notion of black-box strategies captures this interpretation of mixed actions. According to the other viewpoint, players do not randomize, and each player chooses some definite action, but the other players need not know which one, and the mixture represents their uncertainty, i.e. their conjecture about her choice. Aumann and Brandenburger (1995) establish mixed action Nash equilibrium as an equilibrium in conjectures provided they satisfy certain epistemic conditions regarding the common knowledge amongst the players.

In the absence of the betweenness condition, these two viewpoints give rise to different notions of Nash equilibria. Throughout we assume that the player set and their corresponding action sets and payoff functions, as well as the rationality of each player, are common knowledge. A player is said to be rational if, given her beliefs and her preferences, she does not play any suboptimal strategy. Suppose each player plays a fixed action, and these fixed actions are common knowledge, then we get back the notion of pure Nash equilibrium (see definition 3.2). If each player plays a fixed action, but the other players have mixed conjectures over her action, and these conjectures are common knowledge, then this gives us mixed action Nash equilibrium (see definition 3.4). This coincides with the notion of Nash equilibrium as defined in Keskin (2016) and studied further in Phade and Anantharam (2019). Now suppose each player can randomize over her actions and hence implement a black-box strategy. If each player plays a fixed black-box strategy and these black-box strategies are common knowledge, then this gives rise to a new notion of equilibrium. We call it black-box strategy Nash equilibrium (see definition 3.8). If each player plays a fixed black-box strategy and the other players have mixed conjectures over her black-box strategy, and these conjectures are common knowledge, then we get the notion of mixed black-box strategy Nash equilibrium (see definition 3.10).

In the setting of an nn-player normal form game with real valued payoff functions, the pure Nash equilibria do not depend on the specific CPT features of the players, i.e. the reference point, the value function and the two probability weighting functions, one for gains and one for losses. Hence the traditional result on the lack of guarantee for the existence of a pure Nash equilibrium continues to hold when players have CPT preferences. Keskin (2016) proves the existence of a mixed action Nash equilibrium for any finite game when players have CPT preferences. In example 3.9, we show that a finite game may not have any black-box strategy Nash equilibrium. On the other hand, in theorem 3.12, we prove our main result that for any finite game with players having CPT preferences, there exists a mixed black-box strategy Nash equilibrium. If the players have EUT preferences, then the notions of black-box strategy Nash equilibrium and mixed black-box strategy Nash equilibrium are equivalent to the notion of mixed action Nash equilibrium (when interpreted appropriately; see the remark before proposition 3.14; see also figure 6).

The paper is organized as follows. In section 2, we describe the CPT setup and establish that under this setup betweenness is equivalent to independence (theorem 2.3). In section 3, we describe an nn-player non-cooperative game setup and define various notions of Nash equilibrium in the absence of betweenness, in particular with CPT preferences. We discuss the questions concerning their existence and how these different notions of equilibria compare with each other. In section 4, we conclude with a table that summarizes the results.

To close this section, we introduce some notational conventions that will be used in the document. If ZZ is a Polish space (complete separable metric space), let 𝒫(Z)\mathscr{P}(Z) denote the set of all probability measures on (Z,)(Z,\mathscr{F}), where \mathscr{F} is the Borel sigma-algebra of ZZ. Let supp(p)\mathop{\rm supp}(p) denote the support of a distribution p𝒫(Z)p\in\mathscr{P}(Z), i.e. the smallest closed subset of ZZ such that p(Z)=1p(Z)=1. Let Δf(Z)𝒫(Z)\Delta_{f}(Z)\subset\mathscr{P}(Z) denote the set of all probability distributions that have a finite support. For any element pΔf(Z)p\in\Delta_{f}(Z), let p[z]p[z] denote the probability of zZz\in Z assigned by pp. For zZz\in Z, let 1{z}Δf(Z)\textbf{1}\{z\}\in\Delta_{f}(Z) denote the probability distribution such that p[z]=1p[z]=1. If ZZ is finite (and hence a Polish space with respect to the discrete topology), let Δ(Z)\Delta(Z) denote the set of all probability distributions on the set ZZ, viz.

Δ(Z)=𝒫(Z)=Δf(Z)={(p[z])zZ|p[z]0zZ,zZp[z]=1},\Delta(Z)=\mathscr{P}(Z)=\Delta_{f}(Z)=\left\{(p[z])_{z\in Z}\bigg{|}p[z]\geq 0\;\forall z\in Z,\sum_{z\in Z}p[z]=1\right\},

with the usual topology. Let Δm1\Delta^{m-1} denote the standard (m1)(m-1)-simplex, i.e. Δ({1,,m})\Delta(\{1,\dots,m\}). If ZZ is a subset of a Euclidean space, then let co(Z)co(Z) denote the convex hull of ZZ, and let co¯(Z)\overline{co}(Z) denote the closed convex hull of ZZ.

2 Cumulative Prospect Theory and Betweenness

We first describe the setup for CPT (for more details see Wakker (2010)). Each person is associated with a reference point rr\in\mathbb{R}, a value function v:v:\mathbb{R}\to\mathbb{R}, and two probability weighting functions w±:[0,1][0,1]w^{\pm}:[0,1]\to[0,1], w+w^{+} for gains and ww^{-} for losses. The function v(x)v(x) satisfies: (i) it is continuous in xx, (ii) v(r)=0v(r)=0, (iii) it is strictly increasing in xx. The value function is generally assumed to be convex in the losses frame (x<rx<r) and concave in the gains frame (xrx\geq r), and to be steeper in the losses frame than in the gains frame in the sense that v(rz)v(r+z)v(r-z)\leq-v(r+z) for all z0z\geq 0. However, these assumptions are not needed for the results in this paper to hold. The probability weighting functions w±:[0,1][0,1]w^{\pm}:[0,1]\to[0,1] satisfy: (i) they are continuous, (ii) they are strictly increasing, (iii) w±(0)=0w^{\pm}(0)=0and w±(1)=1w^{\pm}(1)=1. We say that (r,v,w±)(r,v,w^{\pm}) are the CPT features of that person.

Suppose a person faces a lottery (or prospect) L:={(pk,zk)}1kmL:=\{(p_{k},z_{k})\}_{1\leq k\leq m}, where zk,1kmz_{k}\in\mathbb{R},1\leq k\leq m, denotes an outcome and pk,1kmp_{k},1\leq k\leq m, is the probability with which outcome zkz_{k} occurs. We assume that k=1mpk=1\sum_{k=1}^{m}p_{k}=1. (Note that we are allowed to have pk=0p_{k}=0 for some values of kk, and we can have zk=zkz_{k}=z_{k^{\prime}} even when kkk\neq k^{\prime}.) Let z:=(zk)1kmz:=(z_{k})_{1\leq k\leq m} and p:=(pk)1kmp:=(p_{k})_{1\leq k\leq m}. We denote LL as (p,z)(p,z) and refer to the vector zz as an outcome profile and pp as a probability vector.

Let α:=(α1,,αm)\alpha:=(\alpha_{1},\dots,\alpha_{m}) be a permutation of (1,,m)(1,\dots,m) such that

zα1zα2zαm.z_{\alpha_{1}}\geq z_{\alpha_{2}}\geq\dots\geq z_{\alpha_{m}}. (2.1)

Let 0krm0\leq k_{r}\leq m be such that zαkrz_{\alpha_{k}}\geq r for 1kkr1\leq k\leq k_{r} and zαk<rz_{\alpha_{k}}<r for kr<kmk_{r}<k\leq m. (Here kr=0k_{r}=0 when zαk<rz_{\alpha_{k}}<r for all 1km1\leq k\leq m.) The CPT value V(L)V(L) of the prospect LL is evaluated using the value function v()v(\cdot) and the probability weighting functions w±()w^{\pm}(\cdot) as follows:

V(L):=k=1krπk+(p,α)v(zαk)+k=kr+1mπk(p,α)v(zαk),V(L):=\sum_{k=1}^{k_{r}}\pi_{k}^{+}(p,\alpha)v(z_{\alpha_{k}})+\sum_{k=k_{r}+1}^{m}\pi_{k}^{-}(p,\alpha)v(z_{\alpha_{k}}), (2.2)

where πk+(p,α),1kkr,\pi^{+}_{k}(p,\alpha),1\leq k\leq k_{r}, and πk(p,α),kr<km,\pi^{-}_{k}(p,\alpha),k_{r}<k\leq m, are decision weights defined via:

π1+(p,α)\displaystyle\pi^{+}_{1}(p,\alpha) :=w+(pα1),\displaystyle:=w^{+}(p_{\alpha_{1}}),
πk+(p,α)\displaystyle\pi_{k}^{+}(p,\alpha) :=w+(pα1++pαk)w+(pα1++pαk1)\displaystyle:=w^{+}(p_{\alpha_{1}}+\dots+p_{\alpha_{k}})-w^{+}(p_{\alpha_{1}}+\dots+p_{\alpha_{k-1}}) for 1<km,\displaystyle 1<k\leq m,
πk(p,α)\displaystyle\pi_{k}^{-}(p,\alpha) :=w(pαm++pαk)w(pαm++pαk+1)\displaystyle:=w^{-}(p_{\alpha_{m}}+\dots+p_{\alpha_{k}})-w^{-}(p_{\alpha_{m}}+\dots+p_{\alpha_{k+1}}) for 1k<m,\displaystyle 1\leq k<m,
πm(p,α)\displaystyle\pi^{-}_{m}(p,\alpha) :=w(pαm).\displaystyle:=w^{-}(p_{\alpha_{m}}).

Although the expression on the right in equation (2.2) depends on the permutation α\alpha, one can check that the formula evaluates to the same value V(L)V(L) as long as the permutation α\alpha satisfies (2.1). The CPT value in equation (2.2) can equivalently be written as:

V(L)\displaystyle V(L) =k=1kr1w+(i=1kpαi)[v(zαk)v(zαk+1)]\displaystyle=\sum_{k=1}^{k_{r}-1}w^{+}\left(\sum_{i=1}^{k}p_{\alpha_{i}}\right)\left[v(z_{\alpha_{k}})-v(z_{\alpha_{k+1}})\right]
+w+(i=1krpαi)v(zαkr)+w(i=kr+1mpαi)v(zαkr+1)\displaystyle+w^{+}\left(\sum_{i=1}^{k_{r}}p_{\alpha_{i}}\right)v\left(z_{\alpha_{k_{r}}}\right)+w^{-}\left(\sum_{i=k_{r}+1}^{m}p_{\alpha_{i}}\right)v(z_{\alpha_{k_{r}+1}})
+k=kr+1m1w(i=k+1mpαi)[v(zαk+1)v(zαk)].\displaystyle+\sum_{k=k_{r}+1}^{m-1}w^{-}\left(\sum_{i=k+1}^{m}p_{\alpha_{i}}\right)\left[v(z_{\alpha_{k+1}})-v(z_{\alpha_{k}})\right]. (2.3)

A person is said to have CPT preferences if, given a choice between prospect L1L_{1} and prospect L2L_{2}, she chooses the one with higher CPT value.

We now define some axioms for preferences over lotteries. We are interested in “mixtures” of lotteries, i.e. lotteries with other lotteries as outcomes. Consider a (two-stage) compound lottery K:={(qj,Lj)}1jtK:=\{(q^{j},L^{j})\}_{1\leq j\leq t}, where Lj=(pj,zj),1jtL^{j}=(p^{j},z^{j}),1\leq j\leq t, are lotteries over real outcomes and qjq^{j} is the chance of lottery LjL^{j}. We assume that j=1tqj=1\sum_{j=1}^{t}q^{j}=1. A two-stage compound lottery can be reduced to a single-stage lottery by multiplying the probability vector pjp^{j} corresponding to the lottery LjL^{j} by qjq^{j} for each j,1jtj,1\leq j\leq t, and then adding the probabilities of identical outcomes across all the lotteries Lj,1jtL^{j},1\leq j\leq t. Let j=1tqjLj\sum_{j=1}^{t}q^{j}L^{j} denote the reduced lottery corresponding to the compound lottery KK.

Let \preceq denote a preference relation over single-stage lotteries. We assume \preceq to be a weak order, i.e. \preceq is transitive (if L1L2L_{1}\preceq L_{2} and L2L3L_{2}\preceq L_{3}, then L1L3L_{1}\preceq L_{3}) and complete (for all L1,L2L_{1},L_{2}, we have L1L2L_{1}\preceq L_{2} or L2L1L_{2}\preceq L_{1}, where possibly both preferences hold). The additional binary relations ,,\succeq,\sim,\prec and \succ are derived from \preceq in the usual manner. A preference relation \preceq is a CPT preference relation if there exist CPT features (r,v,w±)(r,v,w^{\pm}) such that L1L2L_{1}\preceq L_{2} iff V(L1)V(L2)V(L_{1})\leq V(L_{2}). Note that a CPT preference relation is a weak order. A preference relation \preceq satisfies independence if for any lotteries L1,L2L_{1},L_{2} and LL, and any constant 0α10\leq\alpha\leq 1, L1L2L_{1}\preceq L_{2} implies αL1+(1α)LαL2+(1α)L\alpha L_{1}+(1-\alpha)L\preceq\alpha L_{2}+(1-\alpha)L. A preference relation \preceq satisfies betweenness if for any lotteries L1L2L_{1}\preceq L_{2}, we have L1αL1+(1α)L2L2L_{1}\preceq\alpha L_{1}+(1-\alpha)L_{2}\preceq L_{2}, for all 0α10\leq\alpha\leq 1. A preference relation \preceq satisfies weak betweenness if for any lotteries L1L2L_{1}\sim L_{2}, we have L1αL1+(1α)L2L_{1}\sim\alpha L_{1}+(1-\alpha)L_{2}, for all 0α10\leq\alpha\leq 1.

Suppose a preference relation \preceq satisfies independence. Then L1L2L_{1}\preceq L_{2} implies

L1=αL1+(1α)L1αL1+(1α)L2αL2+(1α)L2=L2.L_{1}=\alpha L_{1}+(1-\alpha)L_{1}\preceq\alpha L_{1}+(1-\alpha)L_{2}\preceq\alpha L_{2}+(1-\alpha)L_{2}=L_{2}.

Thus, if a preference relation satisfies independence, then it satisfies betweenness. Also, if a preference relation satisfies betweenness, then it satifies weak betweenness.

In the following example, we will provide CPT features for Alice so that her preferences agree with those described in section 1. This example also shows that cumulative prospect theory can give rise to preferences that do not satisfy betweenness.

Example 2.1.

Recall that Alice is faced with the following three lotteries:

L1={(0.34,$20,000);(0.66,$0)},\displaystyle L_{1}=\{(0.34,\$20,000);(0.66,\$0)\},
L2={(0.17,$30,000);(0.83,$0)},\displaystyle L_{2}=\{(0.17,\$30,000);(0.83,\$0)\},
L={(0.01,$30,000);(0.32,$20,000);(0.67,$0)}.\displaystyle L=\{(0.01,\$30,000);(0.32,\$20,000);(0.67,\$0)\}.

Let r=0r=0 be the reference point of Alice. Thus all the outcomes lie in the gains domain. Let v(x)=x0.8v(x)=x^{0.8} for x0x\geq 0; Alice is risk-averse in the gains domain. Let the probability weighting function for gains be given by

w+(p)=exp{(lnp)0.6},w^{+}(p)=\exp\{-(-\ln p)^{0.6}\},

a form suggested by Prelec (1998) (see figure 1). We won’t need the probability weighting function for losses. Direct computations show that V(L1)=968.96,V(L2)=932.29V(L_{1})=968.96,V(L_{2})=932.29, and V(L)=1022.51V(L)=1022.51 (all decimal numbers in this example are correct to two decimal places). Thus the preference behavior of Alice, as described in section 1 (i.e., she prefers L1L_{1} over L2L_{2}, but prefers LL over L1L_{1} and L2L_{2}), is consistent with CPT and can be modeled, for example, with the CPT features stated here. ∎

The following example shows that CPT can give rise to preferences that do not satisfy weak betweenness (the lotteries and the CPT features presented below appear in Keskin (2016)).

Example 2.2.

Suppose Charlie has r=0r=0 as his reference point and v(x)=xv(x)=x as his value function. Let his probability weighting function for gains be given by

w+(p)=exp{(lnp)0.5}.w^{+}(p)=\exp\{-(-\ln p)^{0.5}\}.

(See figure 1.) We won’t need the probability weighting function for losses since we consider only outcomes in the gains domain in this example. Consider the lotteries L1={(0.5,2β);(0.5,0)}L_{1}=\{(0.5,2\beta);(0.5,0)\} and L2={(0.5,β+1);(0.5,1)}L_{2}=\{(0.5,\beta+1);(0.5,1)\}, where β=1/w+(0.5)=2.299\beta=1/w^{+}(0.5)=2.299 (all decimal numbers in this example are correct to three decimal places). Direct computations reveal that V(L1)=V(L2)=2.000>V(0.5L1+0.5L2)=1.985V(L_{1})=V(L_{2})=2.000>V(0.5L_{1}+0.5L_{2})=1.985.

00.20.20.40.40.60.60.80.81100.20.20.40.40.60.60.80.811ppw+(p)w^{+}(p)
Figure 1: The solid curve shows the probability weighting function for Alice from example 2.1 and example 3.7, and the dashed curve shows the probability weighting function for Charlie from example 2.2.

Given a utility function u:u:\mathbb{R}\to\mathbb{R} (assumed to be continuous and strictly increasing), the expected utility of a lottery L={(pk,zk)}1kmL=\{(p_{k},z_{k})\}_{1\leq k\leq m} is defined as U(L):=k=1mpku(zk)U(L):=\sum_{k=1}^{m}p_{k}u(z_{k}). A preference relation \preceq is said to be an EUT preference relation if there exists a utility function uu such that L1L2L_{1}\preceq L_{2} iff U(L1)U(L2)U(L_{1})\leq U(L_{2}). Note that if the CPT probability weighting functions are linear, i.e. w±(p)=pw^{\pm}(p)=p for 0p10\leq p\leq 1, then the CPT value of a lottery coincides with the expected utility of that lottery with respect to the utility function u=vu=v. It is well known that EUT preference relations satisfy independence and hence betweenness. Several generalizations of EUT have been obtained by weakening the independence axiom and assuming only betweenness, for example, weighted utility theory (Chew and MacCrimmon, 1979; Chew, 1983), skew-symmetric bilinear utility (Fishburn, 1988; Bordley and Hazen, 1991), implicit expected utility (Dekel, 1986; Chew, 1989) and disappointment aversion theory (Gul, 1991; Bordley, 1992). The following theorem shows that in the restricted setting of CPT preferences, betweenness and independence are equivalent.

Theorem 2.3.

If \preceq is a CPT preference relation, then the following are equivalent:

  1. (i)

    \preceq is an EUT preference relation,

  2. (ii)

    \preceq satisfies independence,

  3. (iii)

    \preceq satisfies betweenness.

Proof.

Let the CPT preference relation \preceq be given by (r,v,w±)(r,v,w^{\pm}). Since an EUT preference relation satisfies independence, we get that (i) implies (ii). Since betweenness is a weaker condition than independence, we get that (ii) implies (iii). We will now show that if \preceq satisfies betweenness, then the probability weighting functions are linear, i.e. w±(p)=pw^{\pm}(p)=p for 0p10\leq p\leq 1. This will imply that \preceq is an EUT preference relation with utility function u=vu=v, and hence complete the proof.

Assume that the CPT preference relation \preceq satisfies betweenness. Consider a lottery A:={(p1,z1),(p2,z2),(1p1p2,r)}A:=\{(p_{1},z_{1}),(p_{2},z_{2}),(1-p_{1}-p_{2},r)\} such that z1z2rz_{1}\geq z_{2}\geq r, p10p_{1}\geq 0, p2>0p_{2}>0 and p1+p21p_{1}+p_{2}\leq 1. By (2), we have

V(A)=δ1w+(P1)+δ2w+(P2),V(A)=\delta_{1}w^{+}(P_{1})+\delta_{2}w^{+}(P_{2}),

where δ1:=v(z1)v(z2)\delta_{1}:=v(z_{1})-v(z_{2}), δ2:=v(z2)\delta_{2}:=v(z_{2}), P1:=p1P_{1}:=p_{1} and P2:=p1+p2P_{2}:=p_{1}+p_{2}. Let lottery B:={(q1,z1),(q2,z2),(1q1q2,r)}B:=\{(q_{1},z_{1}),(q_{2},z_{2}),(1-q_{1}-q_{2},r)\} be such that q1,q20q_{1},q_{2}\geq 0, Q1:=q1>p1Q_{1}:=q_{1}>p_{1}, and Q2:=q1+q2<P2Q_{2}:=q_{1}+q_{2}<P_{2}. By (2), we have

V(B)=δ1w+(Q1)+δ2w+(Q2).V(B)=\delta_{1}w^{+}(Q_{1})+\delta_{2}w^{+}(Q_{2}).

If z1,z2,p1,p2,q1z_{1},z_{2},p_{1},p_{2},q_{1} and q2q_{2} are such that

δ1δ2=w+(P2)w+(Q2)w+(Q1)w+(P1),\frac{\delta_{1}}{\delta_{2}}=\frac{w^{+}(P_{2})-w^{+}(Q_{2})}{w^{+}(Q_{1})-w^{+}(P_{1})}, (2.4)

then V(A)=V(B)V(A)=V(B) and, by betweenness, for any 0α10\leq\alpha\leq 1 we have V(A)=V(B)=V(αA+(1α)B)V(A)=V(B)=V(\alpha A+(1-\alpha)B), i.e.

δ1w+(Q1)+δ2w+(Q2)=δ1w+(αP1+(1α)Q1)+δ2w+(αP2+(1α)Q2).\delta_{1}w^{+}(Q_{1})+\delta_{2}w^{+}(Q_{2})=\delta_{1}w^{+}(\alpha P_{1}+(1-\alpha)Q_{1})+\delta_{2}w^{+}(\alpha P_{2}+(1-\alpha)Q_{2}).

Using (2.4) we get

[w+(P2)w+(Q2)]\displaystyle\left[w^{+}(P_{2})-w^{+}(Q_{2})\right] [w+(Q1)w+(αP1+(1α)Q1))]\displaystyle\left[w^{+}(Q_{1})-w^{+}(\alpha P_{1}+(1-\alpha)Q_{1}))\right]
=[w+(Q1)w+(P1)][w+(αP2+(1α)Q2)w+(Q2)].\displaystyle=\left[w^{+}(Q_{1})-w^{+}(P_{1})\right]\left[w^{+}(\alpha P_{2}+(1-\alpha)Q_{2})-w^{+}(Q_{2})\right]. (2.5)

Given any 0P1<Q1Q2<P210\leq P_{1}<Q_{1}\leq Q_{2}<P_{2}\leq 1, there exist z1z_{1} and z2z_{2} such that (2.4) holds. Indeed, take any δ>0\delta>0 belonging to the range of the function vv. This exists because v(r)=0v(r)=0 and vv is a strictly increasing function. Since w+w^{+} is a strictly increasing function, we have

κ:=w+(P2)w+(Q2)w+(Q1)w+(P1)>0.\kappa:=\frac{w^{+}(P_{2})-w^{+}(Q_{2})}{w^{+}(Q_{1})-w^{+}(P_{1})}>0.

Take z2=v1(δ/(1+κ))z_{2}=v^{-1}({\delta}/{(1+\kappa)}) and z1=v1(δ)z_{1}=v^{-1}(\delta). These are well defined because vv is assumed to be continuous and strictly increasing, and δ\delta belongs to its range. Hence z1>z2>rz_{1}>z_{2}>r as required. Thus (2) holds for any 0P1<Q1Q2<P210\leq P_{1}<Q_{1}\leq Q_{2}<P_{2}\leq 1. In particular, when Q1=Q2Q_{1}=Q_{2}, we have

[w+(P2)w+(Q)]\displaystyle\left[w^{+}(P_{2})-w^{+}(Q)\right] [w+(Q)w+(R1)]=[w+(Q)w+(P1)][w+(R2)w+(Q)],\displaystyle\left[w^{+}(Q)-w^{+}(R_{1})\right]=\left[w^{+}(Q)-w^{+}(P_{1})\right]\left[w^{+}(R_{2})-w^{+}(Q)\right],

where Q:=Q1=Q2Q:=Q_{1}=Q_{2}, R1:=αP1+(1α)QR_{1}:=\alpha P_{1}+(1-\alpha)Q and R2:=αP2+(1α)QR_{2}:=\alpha P_{2}+(1-\alpha)Q. Equivalently, for any 0a1<c1<b<c2<a210\leq a_{1}<c_{1}<b<c_{2}<a_{2}\leq 1 such that (a2b)(bc1)=(ba1)(c2b)(a_{2}-b)(b-c_{1})=(b-a_{1})(c_{2}-b), we have

[w+(a2)w+(b)]\displaystyle\left[w^{+}(a_{2})-w^{+}(b)\right] [w+(b)w+(c1)]=[w+(b)w+(a1)][w+(c2)w+(b)].\displaystyle\left[w^{+}(b)-w^{+}(c_{1})\right]=\left[w^{+}(b)-w^{+}(a_{1})\right]\left[w^{+}(c_{2})-w^{+}(b)\right].

In lemma B.1, we prove that the above condition implies w+(p)=pw^{+}(p)=p, for 0p10\leq p\leq 1. Similarly, we can show that w(p)=pw^{-}(p)=p, for 0p10\leq p\leq 1. This completes the proof. ∎

3 Equilibrium in black-box strategies

We now consider an nn-player non-cooperative game where the players have CPT preferences. We will discuss several notions of equilibrium for such a game and will contrast them.

Let Γ:=(N,(Ai)iN,(xi)iN)\Gamma:=(N,(A_{i})_{i\in N},(x_{i})_{i\in N}) denote a game, where N:={1,,n}N:=\{1,\dots,n\} is the set of players, AiA_{i} is the finite action set of player ii, and xi:Ax_{i}:A\to\mathbb{R} is the payoff function for player ii. Here A:=iAiA:=\prod_{i}A_{i} denotes the set of all action profiles a:=(a1,,an)a:=(a_{1},\dots,a_{n}). Let Ai:=ijAjA_{-i}:=\prod_{i\neq j}A_{j} denote the set of action profiles aia_{-i} of all players except player ii.

Definition 3.1.

For any action profile aiAia_{-i}\in A_{-i} of the opponents, we define the best response action set of player ii to be

𝒜i(ai):=argmaxaiAixi(ai,ai).\mathscr{A}_{i}(a_{-i}):=\operatorname*{arg\,max}_{a_{i}\in A_{i}}x_{i}(a_{i},a_{-i}). (3.1)
Definition 3.2.

An action profile a=(a1,,an)a=(a_{1},\dots,a_{n}) is said to be a pure Nash equilibrium if for each player iNi\in N, we have

ai𝒜i(ai).a_{i}\in\mathscr{A}_{i}(a_{-i}).

The notion of pure Nash equilibrium is the same whether the players have CPT preferences or EUT preferences because only deterministic lotteries, comprised of being offered one outcome with probability 11, are considered in the framework of this notion. It is well known that for any given game Γ\Gamma, a pure Nash equilibrium need not exist.

Let μiΔ(Ai)\mu_{-i}\in\Delta(A_{-i}) denote a belief of player ii on the action profiles of her opponents. Given the belief μi\mu_{-i} of player ii, if she decides to play action aia_{i}, then she will face the lottery {(μi[ai],xi(ai,ai))}aiAi.\{(\mu_{-i}[a_{-i}],x_{i}(a_{i},a_{-i}))\}_{a_{-i}\in A_{-i}}.

Definition 3.3.

For any belief μiΔ(Ai)\mu_{-i}\in\Delta(A_{-i}), define the best response action set of player ii as

𝒜i(μi):=argmaxaiAiVi({(μi[ai],xi(ai,ai))}aiAi).\displaystyle\mathscr{A}_{i}(\mu_{-i}):=\operatorname*{arg\,max}_{a_{i}\in A_{i}}V_{i}\left(\left\{\left(\mu_{-i}[a_{-i}],x_{i}(a_{i},a_{-i})\right)\right\}_{a_{-i}\in A_{-i}}\right). (3.2)

Note that this definition is consistent with the definition of the best response action set that takes an action profile aia_{-i} of the opponents as its input (definition 3.1), if we interpret aia_{-i} as the belief 1{ai}Δ(Ai)\textbf{1}\{a_{-i}\}\in\Delta(A_{-i}), since 𝒜i(1{ai})=𝒜i(ai)\mathscr{A}_{i}(\textbf{1}\{a_{-i}\})=\mathscr{A}_{i}(a_{-i}).

Let σiΔ(Ai)\sigma_{i}\in\Delta(A_{i}) denote a conjecture over the action of player ii. Let σ:=(σ1,,σn)\sigma:=(\sigma_{1},\dots,\sigma_{n}) denote a profile of conjectures, and let σi:=(σj)ji\sigma_{-i}:=(\sigma_{j})_{j\neq i} denote the profile of conjectures for all players except player ii. Let μi(σi)Δ(Ai)\mu_{-i}(\sigma_{-i})\in\Delta(A_{-i}) be the belief induced by conjectures σj,ji\sigma_{j},j\neq i, given by

μi(σi)[ai]:=jiσj[ai],\mu_{-i}(\sigma_{-i})[a_{-i}]:=\prod_{j\neq i}\sigma_{j}[a_{-i}],

which is nothing but the product distribution induced by σi\sigma_{-i}.

Definition 3.4.

A conjecture profile σ=(σ1,,σn)\sigma=(\sigma_{1},\dots,\sigma_{n}) is said to be a mixed action Nash equilibrium if, for each player ii, we have

ai𝒜i(μi(σi)), for all aisuppσi.a_{i}\in\mathscr{A}_{i}(\mu_{-i}(\sigma_{-i})),\text{ for all }a_{i}\in\mathop{\rm supp}\sigma_{i}.

In other words, the conjecture σi\sigma_{i} over the action of player ii should assign positive probabilities to only optimal actions of player ii, given her belief μi(σi)\mu_{-i}(\sigma_{-i}).

It is well known that a mixed Nash equilibrium exists for every game with EUT players, see Nash (1951). Keskin (2016) generalizes the result of Nash (1951) on the existence of a mixed action Nash equilibrium to the case when players have CPT preferences.

Let Bi:=Δ(Ai)B_{i}:=\Delta(A_{i}) denote the set of all black-box strategies for player ii with a typical element denoted by biBib_{i}\in B_{i}. Recall that if player ii implements a black-box strategy bib_{i}, then we interpret this as a trusted party other than the player sampling an action aiAia_{i}\in A_{i} from the distribution bib_{i} and playing action aia_{i} on behalf of player ii. We assume the usual topology on BiB_{i}. Let B:=iBiB:=\prod_{i}B_{i} and Bi:=jiBjB_{-i}:=\prod_{j\neq i}B_{j} with typical elements denoted by bb and bib_{-i}, respectively.

Note that, although a conjecture σi\sigma_{i} and a black-box strategy bib_{i} are mathematically equivalent, viz. they are elements of the same set Bi=Δ(Ai)B_{i}=\Delta(A_{i}), they have different interpretations. We will call siΔ(Ai)s_{i}\in\Delta(A_{i}) a mixture of actions of player ii when we want to be agnostic to which interpretation is being imposed. Let Si:=Δ(Ai),S:=iΔ(Ai)S_{i}:=\Delta(A_{i}),S:=\prod_{i}\Delta(A_{i}) and Si:=jiSiS_{-i}:=\prod_{j\neq i}S_{i} with typical elements denoted by si,ss_{i},s and sis_{-i}, respectively. (Note that SΔ(A)S\neq\Delta(A) unless all but one player have singleton action sets.)

For any belief μiΔ(Ai)\mu_{-i}\in\Delta(A_{-i}) and any black-box strategy bib_{i} of player ii, let μ(bi,μi)Δ(A)\mu(b_{i},\mu_{-i})\in\Delta(A) denote the product distribution given by

μ(bi,μi)[a]:=bi[ai]μi[ai].\mu(b_{i},\mu_{-i})[a]:=b_{i}[a_{i}]\mu_{-i}[a_{-i}].

Given the belief μi\mu_{-i} of player ii, if she decides to implement the black-box strategy bib_{i}, then she will face the lottery {μ(bi,μi)[a],xi(a))}aA\{\mu(b_{i},\mu_{-i})[a],x_{i}(a))\}_{a\in A}.

Definition 3.5.

For any belief μiΔ(Ai)\mu_{-i}\in\Delta(A_{-i}), define the best response black-box strategy set of player ii as

i(μi):=argmaxbiBiVi({(μ(bi,μi)[a],xi(a))}aA).\displaystyle\mathscr{B}_{i}(\mu_{-i}):=\operatorname*{arg\,max}_{b_{i}\in B_{i}}V_{i}\left(\left\{\left(\mu(b_{i},\mu_{-i})[a],x_{i}(a)\right)\right\}_{a\in A}\right).
Lemma 3.6.

For any belief μi\mu_{-i}, the set i(μi)\mathscr{B}_{i}(\mu_{-i}) is non-empty, and

co¯(i(μi))=co(i(μi)).\overline{co}(\mathscr{B}_{i}(\mu_{-i}))=co(\mathscr{B}_{i}(\mu_{-i})).
Proof.

For a lottery L=(p,z)L=(p,z), where z=(zk)1kmz=(z_{k})_{1\leq k\leq m} is the outcome profile, and (pk)1km(p_{k})_{1\leq k\leq m} is the probability vector, the function Vi(p,z)V_{i}(p,z) is continuous with respect to pΔm1p\in\Delta^{m-1} (Keskin, 2016). Thus, Vi({(μ(bi,μi)[a],xi(a))}aA)V_{i}(\{(\mu(b_{i},\mu_{-i})[a],x_{i}(a))\}_{a\in A}) is a function continuous with respect to biBib_{i}\in B_{i}, and hence i(μi)\mathscr{B}_{i}(\mu_{-i}) is a non-empty closed subset of the compact space BiB_{i}. Since the convex hull of a compact subset of a Euclidean space is compact (Rudin, 1991, Chapter 3), the set co(i(μi))co(\mathscr{B}_{i}(\mu_{-i})) is closed. This completes the proof. ∎

Let us compare the two concepts: the best response action set (definition 3.3) and the best response black-box strategy set (definition 3.5). Even though both of them take the belief μi\mu_{-i} of player ii as input, the best response action set 𝒜i(μi)\mathscr{A}_{i}(\mu_{-i}) outputs a collection of actions of player ii, whereas the best response black-box strategy set i(μi)\mathscr{B}_{i}(\mu_{-i}) outputs a collection of black-box strategies of player ii, which are probability distributions over the set of actions aiAia_{i}\in A_{i}. If we interpret an action aia_{i} as the mixture 1{ai}Si=Δ(Ai)\textbf{1}\{a_{i}\}\in S_{i}=\Delta(A_{i}), and a black-box strategy bib_{i} as a mixture as well, then we can compare the two sets 𝒜(μi)\mathscr{A}(\mu_{-i}) and (μi)\mathscr{B}(\mu_{-i}) as subsets of SiS_{i}. The following example shows that, in general, the two sets can be disjoint, and hence quite distinct.

Example 3.7.

We consider a 22-player game. Let Alice be player 11, with action set A1={1,2}A_{1}=\{1,2\}, and let Bob be player 22, with action set A2={1,2,3}A_{2}=\{1,2,3\}. Let the payoff function for Alice be as shown in figure 2. Let μ1=(0.17,0.17,0.66)Δ(A1)=Δ(A2)\mu_{-1}=(0.17,0.17,0.66)\in\Delta(A_{-1})=\Delta(A_{2}) be the belief of Alice. Then, as considered in section 1, Alice faces the lottery L1={(0.34,$20,000);(0.66,$0)}L_{1}=\{(0.34,\$20,000);(0.66,\$0)\} if she plays action 11 and the lottery L2={(0.17,$30,000);(0.83,$0)}L_{2}=\{(0.17,\$30,000);(0.83,\$0)\} if she plays action 22. We retain the CPT features for Alice, as in example 2.1, viz.: r=0r=0, v(x)=x0.8v(x)=x^{0.8} for x0x\geq 0, and

w+(p)=exp{(lnp)0.6}.w^{+}(p)=\exp\{-(-\ln p)^{0.6}\}.

We saw that V1(L1)=968.96V_{1}(L_{1})=968.96, V1(L2)=932.29V_{1}(L_{2})=932.29, and V(16/17L1+1/17L2)=1022.51V(16/17L_{1}+1/17L_{2})=1022.51 (all decimal numbers in this example are correct to two decimal places). Amongst all the mixtures, the maximum CPT value is achieved at the unique mixture L=αL1+(1α)L2L^{*}=\alpha^{*}L_{1}+(1-\alpha^{*})L_{2}, where α=0.96\alpha^{*}=0.96; we have V1(L)=1023.16V_{1}(L^{*})=1023.16. Thus, 𝒜1(μ1)={1{1}}\mathscr{A}_{1}(\mu_{-1})=\{\textbf{1}\{1\}\} and 1(μ1)={(α,1α)}\mathscr{B}_{1}(\mu_{-1})=\{(\alpha^{*},1-\alpha^{*})\}.

11 22 33
11 $20,000\$20{,}000 $20,000\$20{,}000 $0\$0
22 $30,000\$30{,}000 $0\$0 $0\$0
Figure 2: Payoff matrix for Alice in example 3.7. Rows and columns correspond to Alice’s and Bob’s actions respectively. The amount in each cell corresponds to Alice’s payoff.

For any black-box strategy profile bib_{-i} of the opponents, let μi(bi)Δ(Ai)\mu_{-i}(b_{-i})\in\Delta(A_{-i}) be the induced belief given by

μi(bi)[ai]:=jibj[ai].\mu_{-i}(b_{-i})[a_{-i}]:=\prod_{j\neq i}b_{j}[a_{-i}].
Definition 3.8.

A black-box strategy profile b=(b1,,bn)b=(b_{1},\dots,b_{n}) is said to be a black-box strategy Nash equilibrium if, for each player ii, we have

bii(μi(bi)).b_{i}\in\mathscr{B}_{i}(\mu_{-i}(b_{-i})).

If the players have EUT preferences, a conjecture profile σ=(σ1,,σn)\sigma=(\sigma_{1},\dots,\sigma_{n}) is a mixed action Nash equilibrium if and only if the black-box strategy profile b=(b1,,bn)b=(b_{1},\dots,b_{n}), where bi=σib_{i}=\sigma_{i}, for all iNi\in N, is a black-box strategy Nash equilibrium. Thus, under EUT, the notion of a black-box strategy Nash equilibrium is equivalent to the notion of a mixed action Nash equilibrium, although there is still a conceptual difference between these two notions based on the interpretations for the mixtures of actions. Further, we have the existence of a black-box strategy Nash equilibrium for any game when players have EUT preferences from the well-known result about the existence of a mixed action Nash equilibrium. The following example shows that, in general, a black-box strategy Nash equilibrium may not exist when players have CPT preferences.

Example 3.9.

Consider a 2×22\times 2 game (i.e a 22-player game where each player has two actions {0,1}\{0,1\}) with the payoff matrices as shown in figure 3. Let the reference points be r1=r2=0r_{1}=r_{2}=0. Let vi()v_{i}(\cdot) be the identity function for i=1,2i=1,2. Let the probability weighting functions for gains for the two players be given by

wi+(p)=exp{(lnp)γi}, for i=1,2,w_{i}^{+}(p)=\exp\{-(-\ln p)^{\gamma_{i}}\},\text{ for }i=1,2,

where γ1=0.5\gamma_{1}=0.5 and γ2=1\gamma_{2}=1. We do not need the probability weighting functions for losses since all the outcomes lie in the gains domain for both the players. Notice that player 22 has EUT preferences since w2+(p)=pw_{2}^{+}(p)=p.

0 1
0 44 0
1 33 11
0 1
0 0 11
1 11 0
Figure 3: Payoff matrices for the 2×22\times 2 game in example 3.9 (left matrix for player 11 and right matrix for player 22). The rows and the columns correspond to the actions of player 11 and player 22, respectively, and the entries in the cell represent the corresponding payoffs.

Suppose player 11 and player 22 play black-box strategies (1p,p)(1-p,p) and (1q,q)(1-q,q), respectively, where p,q[0,1]p,q\in[0,1]. With an abuse of notation, we identify these black-box strategies by pp and qq, respectively. The corresponding lottery faced by player 11 is given by

L1(p,q):={(μ[0,0],4);(μ[1,0],3);(μ[1,1],1);(μ[0,1],0)},L_{1}(p,q):=\{(\mu[0,0],4);(\mu[1,0],3);(\mu[1,1],1);(\mu[0,1],0)\},

where μ[0,0]:=(1p)(1q),μ[1,0]:=p(1q),μ[0,1]:=(1p)q\mu[0,0]:=(1-p)(1-q),\mu[1,0]:=p(1-q),\mu[0,1]:=(1-p)q, and μ[1,1]:=pq\mu[1,1]:=pq. By (2.2), the CPT value of the lottery faced by player 11 is given by

V1(L1(p,q))\displaystyle V_{1}(L_{1}(p,q)) :=4×[w1+(μ[0,0])]\displaystyle:=4\times\left[w^{+}_{1}(\mu[0,0])\right]
+3×[w1+(μ[0,0]+μ[1,0])w1+(μ[0,0]))]\displaystyle+3\times\left[w^{+}_{1}(\mu[0,0]+\mu[1,0])-w^{+}_{1}(\mu[0,0]))\right]
+1×[w1+(μ[0,0]+μ[1,0]+μ[1,1])w1+(μ[0,0]+μ[1,0])].\displaystyle+1\times\left[w^{+}_{1}(\mu[0,0]+\mu[1,0]+\mu[1,1])-w^{+}_{1}(\mu[0,0]+\mu[1,0])\right].

The plot of the function V1(L1(p,q))V_{1}(L_{1}(p,q)) with respect to pp, for q=0.3q=0.3 and q=0.35q=0.35, is shown in figure 4. We observe that the best response black-box strategy set 1(μ1(q))\mathscr{B}_{1}(\mu_{-1}(q)) of player 11 to player 22’s black-box strategy qB2q\in B_{2} satisfies the following: 1(μ1(q))={0}\mathscr{B}_{1}(\mu_{-1}(q))=\{0\} for q<qq<q^{*}, 1(μ1(q))={0,p}\mathscr{B}_{1}(\mu_{-1}(q))=\{0,p^{*}\} for q=qq=q^{*}, and 1(μ1(q))[p,1]\mathscr{B}_{1}(\mu_{-1}(q))\subset[p^{*},1] for q>qq>q^{*}, where p=0.996p^{*}=0.996 and q=0.340q^{*}=0.340 (here the numbers are correct to three decimal points). Further, 1(μ1(q))\mathscr{B}_{1}(\mu_{-1}(q)) is singleton for q(q,1]q\in(q^{*},1] and the unique element in 1(μ1(q))\mathscr{B}_{1}(\mu_{-1}(q)) increases monotonically with respect to qq from pp^{*} to 11 (see figure 5). In particular, 1(μ1(1))={1}\mathscr{B}_{1}(\mu_{-1}(1))=\{1\}. The lottery faced by player 22 is given by

L2(p,q):={(μ[0,0],0);(μ[1,0],1);(μ[1,1],0);(μ[0,1],1)},L_{2}(p,q):=\{(\mu[0,0],0);(\mu[1,0],1);(\mu[1,1],0);(\mu[0,1],1)\},

and the CPT value of player 22 for this lottery is given by V2(L2(p,q))=p(1q)+q(1p)V_{2}(L_{2}(p,q))=p(1-q)+q(1-p). The best response black-box strategy set 2(μ2(p))\mathscr{B}_{2}(\mu_{-2}(p)) of player 22 to player 11’s black-box strategy pB1p\in B_{1} satisfies the following: 2(μ2(p))={1}\mathscr{B}_{2}(\mu_{-2}(p))=\{1\} for p<0.5p<0.5, 2(μ2(p))=[0,1]\mathscr{B}_{2}(\mu_{-2}(p))=[0,1] for p=0.5p=0.5, and 2(μ2(p))={0}\mathscr{B}_{2}(\mu_{-2}(p))=\{0\} for p>0.5p>0.5. As a result, see figure 5, there does not exist any (p,q)(p^{\prime},q^{\prime}) such that p1(μ1(q))p^{\prime}\in\mathscr{B}_{1}(\mu_{-1}(q^{\prime})) and q2(μ2(p))q^{\prime}\in\mathscr{B}_{2}(\mu_{-2}(p^{\prime})), and hence no black-box strategy Nash equilibrium exists for this game.

00.10.10.20.20.30.30.40.40.50.50.60.60.70.70.80.80.90.9112.052.052.12.12.152.152.22.2ppV1(L1(p,q))V_{1}(L_{1}(p,q))q=0.3q=0.3q=0.35q=0.35
Figure 4: The CPT value of player 11 in example 3.9. Here, pp and qq denote the black-box strategies for player 11 and 22, respectively. Note the rise and sharp drop in the two curves near p=1p=1. For the curve for q=0.3q=0.3, the global maximum is attained at p=0p=0, whereas, for the curve for q=0.35q=0.35, the global maximum is attained close to p=1p=1, specifically for some p[0.9,1]p\in[0.9,1].
Refer to caption
Figure 5: The figure (not to scale) shows the best response black-box strategy sets of the two players for the game in example 3.9. The red (dashed) line shows the best response black-box strategy set of player 22 in response to the black-box strategy (1p,p)(1-p,p) of player 11. The green (solid) line shows the best response black-box strategy set of player 11 in response to the black-box strategy (1q,q)(1-q,q) of player 22. Note that there is no intersection of these lines.

Let τi𝒫(Bi)\tau_{i}\in\mathscr{P}(B_{i}) denote a conjecture over the black-box strategy of player ii. This will induce a conjecture σi(τi)Δ(Ai)\sigma_{i}(\tau_{i})\in\Delta(A_{i}) over the action of player ii, given by

σi(τi)[ai]=𝔼τibi[ai].\sigma_{i}(\tau_{i})[a_{i}]=\mathbb{E}_{\tau_{i}}b_{i}[a_{i}].

Given conjectures over black-box strategies (τjΔ(Bj),ji)(\tau_{j}\in\Delta(B_{j}),j\neq i), let σi(τi):=(σj(τj))ji\sigma_{-i}(\tau_{-i}):=(\sigma_{j}(\tau_{j}))_{j\neq i}.

Definition 3.10.

A profile of conjectures over black-box strategies τ=(τ1,,τn)\tau=(\tau_{1},\dots,\tau_{n}) is said to be a mixed black-box strategy Nash equilibrium if, for each player ii, we have

bii(μi(σi(τi))), for all bisuppτi.b_{i}\in\mathscr{B}_{i}(\mu_{-i}(\sigma_{-i}(\tau_{-i}))),\text{ for all }b_{i}\in\mathop{\rm supp}\tau_{i}.
Proposition 3.11.

For a profile of conjectures σ=(σ1,,σn)\sigma^{*}=(\sigma_{1}^{*},\dots,\sigma_{n}^{*}), consider the condition

σico¯(i(μi(σi))), for all i.\sigma_{i}^{*}\in\overline{co}(\mathscr{B}_{i}(\mu_{-i}(\sigma_{-i}^{*}))),\text{ for all }i. (3.3)
  1. (i)

    If τ\tau is a mixed black-box strategy Nash equilibrium, then the profile of conjectures σ\sigma^{*}, where σi=σi(τi),i\sigma_{i}^{*}=\sigma_{i}(\tau_{i}),\forall i, satisfies (3.3).

  2. (ii)

    If σ\sigma^{*} satisfies (3.3), then there exists a profile of finite support conjectures on black-box strategies τ˙=(τ˙1,,τ˙n)\dot{\tau}=(\dot{\tau}_{1},\dots,\dot{\tau}_{n}), where τ˙iΔf(Bi),i\dot{\tau}_{i}\in\Delta_{f}(B_{i}),\forall i, that is a mixed black-box strategy Nash equilibrium, such that σi=σi(τ˙i),i\sigma_{i}^{*}=\sigma_{i}(\dot{\tau}_{i}),\forall i.

Proof.

Suppose τ\tau is a mixed black-box strategy Nash equilibrium. Let σi=σi(τi)\sigma^{*}_{i}=\sigma_{i}(\tau_{i}). Then, for all bisuppτib_{i}\in\mathop{\rm supp}\tau_{i}, we have bii(μi(σi))b_{i}\in\mathscr{B}_{i}(\mu_{-i}(\sigma^{*}_{-i})), and hence σico¯(i(μi(σi)))\sigma^{*}_{i}\in\overline{co}(\mathscr{B}_{i}(\mu_{-i}(\sigma^{*}_{-i}))). This proves statement (i).

For statement (ii), suppose σ\sigma^{*} satisfies condition (3.3). In fact, by lemma 3.6 we have, σico(i(μi(σi)))Δ(Ai)\sigma^{*}_{i}\in co(\mathscr{B}_{i}(\mu_{-i}(\sigma^{*}_{-i})))\subset\Delta(A_{i}), and by Caratheodory’s theorem, σi\sigma^{*}_{i} is a convex combination of at most |Ai||A_{i}| elements in i(μi(σi))\mathscr{B}_{i}(\mu_{-i}(\sigma^{*}_{-i})). Hence, we can construct a mixed black-box strategy Nash equilibrium τ˙\dot{\tau} such that τ˙iΔf(Bi)\dot{\tau}_{i}\in\Delta_{f}(B_{i}) and σi=σi(τ˙i),i\sigma_{i}^{*}=\sigma_{i}(\dot{\tau}_{i}),\forall i. ∎

The content of this proposition is that in order to determine whether a profile τ\tau of conjectures on black box strategies is a mixed black-box strategy Nash equilibrium or not it suffices to study the associated profile of conjectures on actions that is induced by τ\tau. This justifies the study of the set mBBNE\mathrm{mBBNE} discussed below.

Theorem 3.12.

For any game Γ\Gamma, there exists a profile of conjectures σ=(σ1,,σn)\sigma^{*}=(\sigma_{1}^{*},\dots,\sigma_{n}^{*}) that satisfies (3.3).

Proof.

The idea is to use the Kakutani fixed-point theorem, as in the proof of the existence of mixed action Nash equilibrium (Nash, 1950). Assume the usual topology on SiS_{i}, for each ii, and let SS have the corresponding product topology. The set SS is a non-empty compact convex subset of the Euclidean space i|Ai|\prod_{i}\mathbb{R}^{|A_{i}|}. Let K(σ)K(\sigma) be the set-valued function given by

K(σ):=ico¯(i(μi(σi))),K(\sigma):=\prod_{i}\overline{co}(\mathscr{B}_{i}(\mu_{-i}(\sigma_{-i}))),

for all σS\sigma\in S. Since co¯(i(μi(σi)))\overline{co}(\mathscr{B}_{i}(\mu_{-i}(\sigma_{-i}))) is non-empty and convex for each ii (lemma 3.6), the function K(σ)K(\sigma) is non-empty and convex for any σS\sigma\in S. We now show that the function K()K(\cdot) has a closed graph. Let {σt}t=1\{\sigma^{t}\}_{t=1}^{\infty} and {st}t=1\{s^{t}\}_{t=1}^{\infty} be two sequences in SS that converge to σ¯\bar{\sigma} and s¯\bar{s}, respectively, and let stK(σt)s^{t}\in K(\sigma^{t}) for all tt. It is enough to show that s¯K(σ¯)\bar{s}\in K(\bar{\sigma}). For all siSi,σiSis_{i}\in S_{i},\sigma_{-i}\in S_{-i}, let

V~i(si,σi):=supτi𝒫(Bi),𝔼τibi=si𝔼τiVi({(μ(bi,μi(σi))[a],xi(a))}aA).\tilde{V}_{i}(s_{i},\sigma_{-i}):=\sup_{\begin{subarray}{c}\tau_{i}\in\mathscr{P}(B_{i}),\\ \mathbb{E}_{\tau_{i}}b_{i}=s_{i}\end{subarray}}\mathbb{E}_{\tau_{i}}V_{i}\left(\{(\mu(b_{i},\mu_{-i}(\sigma_{-i}))[a],x_{i}(a))\}_{a\in A}\right).

Since the product distribution μ(bi,μi(σi))\mu(b_{i},\mu_{-i}(\sigma_{-i})) is jointly continuous in bib_{i} and σi\sigma_{-i}, and, as noted earlier, Vi(p,z)V_{i}(p,z) is continuous with respect to the probability vector pp, for any fixed outcome profile zz, the function Vi({μ(bi,μi(σi))[a],xi(a)}aA)V_{i}\left(\{\mu(b_{i},\mu_{-i}(\sigma_{-i}))[a],x_{i}(a)\}_{a\in A}\right) is jointly continuous in bib_{i} and σi\sigma_{-i}. This implies that the function V~i(si,σi)\tilde{V}_{i}(s_{i},\sigma_{-i}) is jointly continuous in sis_{i} and σi\sigma_{-i} (see Appendix A). From the definition of V~i\tilde{V}_{i}, it follows that

maxsiΔ(Ai)V~i(si,σi)=maxbiBiVi({(μ(bi,μi(σi))[a],xi(a))}aA).\max_{s_{i}\in\Delta(A_{i})}\tilde{V}_{i}(s_{i},\sigma_{-i})=\max_{b_{i}\in B_{i}}V_{i}\left(\{(\mu(b_{i},\mu_{-i}(\sigma_{-i}))[a],x_{i}(a))\}_{a\in A}\right).

Indeed, the maximum on the left-hand side is well-defined since Δ(Ai)\Delta(A_{i}) is a compact space and V~i(,σi)\tilde{V}_{i}(\cdot,\sigma_{-i}) is a continuous function. The maximum on the right-hand side is well-defined and the maximum is achieved by all bii(μi(σi))b_{i}\in\mathscr{B}_{i}(\mu_{-i}(\sigma_{-i})) (lemma 3.6). Hence,

argmaxsiΔ(Ai)V~i(si,σi)=co¯(i(μi(σi))).\operatorname*{arg\,max}_{s_{i}\in\Delta(A_{i})}\tilde{V}_{i}(s_{i},\sigma_{-i})=\overline{co}(\mathscr{B}_{i}(\mu_{-i}(\sigma_{-i}))).

Since sitco¯(i(μi(σ¯it)))s_{i}^{t}\in\overline{co}(\mathscr{B}_{i}(\mu_{-i}(\bar{\sigma}_{-i}^{t}))), for all tt, we have

V~i(sit,σit)V~i(s~i,σit),for all s~iSi.\tilde{V}_{i}(s_{i}^{t},\sigma^{t}_{-i})\geq\tilde{V}_{i}(\tilde{s}_{i},\sigma^{t}_{-i}),\leavevmode\nobreak\ \leavevmode\nobreak\ \mbox{for all $\tilde{s}_{i}\in S_{i}$.}

Since V~i(si,σi)\tilde{V}_{i}(s_{i},\sigma_{-i}) is jointly continuous in sis_{i} and σi\sigma_{-i}, we get

V~i(s¯i,σ¯i)V~i(s~i,σ¯i),for all s~iSi.\tilde{V}_{i}(\bar{s}_{i},\bar{\sigma}_{-i})\geq\tilde{V}_{i}(\tilde{s}_{i},\bar{\sigma}_{-i}),\leavevmode\nobreak\ \leavevmode\nobreak\ \mbox{for all $\tilde{s}_{i}\in S_{i}$.}

Hence we have s¯ico¯(i(μi(σ¯it)))\bar{s}_{i}\in\overline{co}(\mathscr{B}_{i}(\mu_{-i}(\bar{\sigma}_{-i}^{t}))). This shows that the function K()K(\cdot) has a closed graph. By the Kakutani fixed-point theorem, there exists σ\sigma^{*} such that σK(σ)\sigma^{*}\in K(\sigma^{*}), i.e. σ\sigma^{*} satisfies condition (3.3) (Kakutani, 1941). This completes the proof. ∎

Corollary 3.13.

For any finite game Γ\Gamma, there exists a mixed black-box strategy Nash equilibrium. In particular, there is one that is a profile of finite support conjectures over the black-box strategies of players.

Proof.

Follows from theorem 3.12 and statement (ii) of proposition 3.11. ∎

We now compare the different notions of Nash equilibrium defined above. To that end, we will associate each of the equilibrium notions with their corresponding natural profile of mixtures over actions. For example, corresponding to any pure Nash equilibrium a=(a1,,an)a=(a_{1},\dots,a_{n}), assign the profile of mixtures over actions (1{a1},,1{an})S(\textbf{1}\{a_{1}\},\dots,\textbf{1}\{a_{n}\})\in S. Let pNES\mathrm{pNE}\subset S denote the set of all profiles of mixtures over actions that correspond to pure Nash equilibria. Let mNES\mathrm{mNE}\subset S denote the set of all mixed action Nash equilibria σS\sigma\in S. Let BBNES\mathrm{BBNE}\subset S denote the set of all black-box strategy Nash equilibria bSb\in S. Corresponding to any mixed black-box strategy Nash equilibrium τ=(τ1,,τn)\tau=(\tau_{1},\dots,\tau_{n}), assign the profiles of mixtures over actions (σ1(τ1),,σn(τn))S(\sigma_{1}(\tau_{1}),\dots,\sigma_{n}(\tau_{n}))\in S, and let mBBNES\mathrm{mBBNE}\subset S denote the set of all such profiles. Note that each of the above subsets depends on the underlying game Γ\Gamma and the CPT features of the players.

Proposition 3.14.

For any fixed game Γ\Gamma and CPT features of the players, we have

  1. (i)

    pNEmNE,\mathrm{pNE}\subset\mathrm{mNE},

  2. (ii)

    pNEBBNE,\mathrm{pNE}\subset\mathrm{BBNE}, and

  3. (iii)

    BBNEmBBNE.\mathrm{BBNE}\subset\mathrm{mBBNE}.

Proof.

The proof of statement (i) can be found in Keskin (2016).

For statement (ii), let (1{a1},,1{an})pNE(\textbf{1}\{a_{1}\},\dots,\textbf{1}\{a_{n}\})\in\mathrm{pNE}. For a black-box strategy bib_{i} of player ii, the belief μi=1{ai}\mu_{-i}=\textbf{1}\{a_{-i}\} of player ii gives rise to the lottery {(bi[ai],xi(ai,ai))}aiAi\{(b_{i}[a_{i}^{\prime}],x_{i}(a_{i}^{\prime},a_{-i}))\}_{a_{i}^{\prime}\in A_{i}}. From the definition of CPT value (see equation (2.2)), we observe that Vi({(bi[ai],xi(ai,ai))}aiAi)V_{i}(\{(b_{i}[a_{i}^{\prime}],x_{i}(a_{i}^{\prime},a_{-i}))\}_{a_{i}^{\prime}\in A_{i}}) is optimal as long as the probability distribution bib_{i} does not assign positive probability to any suboptimal outcome. Hence,

i(1{ai})=co¯(1{ai}Si:ai𝒜i(1{ai})).\mathscr{B}_{i}(\textbf{1}\{a_{-i}\})=\overline{co}(\textbf{1}\{a_{i}^{\prime}\}\in S_{i}:a_{i}^{\prime}\in\mathscr{A}_{i}(\textbf{1}\{a_{-i}\})).

In particular, 1{ai}i(1{ai})\textbf{1}\{a_{i}\}\in\mathscr{B}_{i}(\textbf{1}\{a_{-i}\}), and hence (1{a1},,1{an})BBNE(\textbf{1}\{a_{1}\},\dots,\textbf{1}\{a_{n}\})\in\mathrm{BBNE}.

Statement (iii) follows directly from the definitions 3.8 and 3.10. ∎

SSpNE\mathrm{pNE}mNE\mathrm{mNE}BBNE\mathrm{BBNE}BBNE\mathrm{BBNE}mBBNE\mathrm{mBBNE}
Figure 6: Venn diagram depicting the different notions of equilibrium as subsets of the set S=iΔ(Ai)S=\prod_{i}\Delta(A_{i}). The sets marked pNE,mNE,BBNE\mathrm{pNE},\mathrm{mNE},\mathrm{BBNE}, and mBBNE\mathrm{mBBNE} represent the sets of pure Nash equilibria, mixed action Nash equilibria, black-box strategy Nash equilibria, and mixed black-box strategy Nash equilibria, respectively. Examples are given in the body of the text of CPT games lying in each of the indicated regions (a) through (g).

In the following, we show via examples that each of the labeled regions ((a)–(g)), in figure 6, is non-empty in general.

Example 3.15.

For each of the seven regions in figure 6, we provide a 2×22\times 2 game with the accompanying CPT features for the two players verifying that the corresponding region is non-empty. Let the action sets be A1=A2={0,1}A_{1}=A_{2}=\{0,1\}. With an abuse of notation, let p,q[0,1]p,q\in[0,1] denote the mixtures over actions for players 11 and 22, respectively, where pp and qq are the probabilities corresponding to action 11 for both the players. Thus, the set of all profiles of mixtures over actions is S={(p,q):p,q[0,1]}S=\{(p,q):p,q\in[0,1]\}. Let L1(p,q)L_{1}(p,q) and L2(p,q)L_{2}(p,q) denote the corresponding lotteries faced by the two players. (All decimal numbers in these examples are correct to three decimal places.)

  1. (a)

    Let both the players have EUT preferences with their utility functions given by the identity functions ui(x)=xu_{i}(x)=x, for i=1,2i=1,2. Let the payoff matrix be as shown in figure 7. Clearly, (p=0,q=0)pNE(p=0,q=0)\in\mathrm{pNE}.

  2. (b)

    Let ri=0,vi(x)=xr_{i}=0,v_{i}(x)=x, for i=1,2i=1,2. Let w1+(p)=p0.5w_{1}^{+}(p)=p^{0.5} and w2+(p)=pw_{2}^{+}(p)=p, for p[0,1]p\in[0,1]. Let the payoff matrix be as shown in figure 7, where β:=1/w1+(0.5)=1.414\beta:=1/w_{1}^{+}(0.5)=1.414. We have

    L1(p,q)={((1p)(1q),2β);(p(1q),β+1);(pq,1);((1p)q,0)}.L_{1}(p,q)=\{((1-p)(1-q),2\beta);(p(1-q),\beta+1);(pq,1);((1-p)q,0)\}.

    The way β\beta is defined, we get V1(L1(0,0.5))=V1(L1(1,0.5))=2V_{1}(L_{1}(0,0.5))=V_{1}(L_{1}(1,0.5))=2. Also, observe that V2(L2(0.5,0))=V2(L2(0.5,q))=V2(L2(0.5,1)),q[0,1].V_{2}(L_{2}(0.5,0))=V_{2}(L_{2}(0.5,q))=V_{2}(L_{2}(0.5,1)),\forall q\in[0,1]. With these observations, we get that (0.5,0.5)mNE(0.5,0.5)\in\mathrm{mNE}. We have, argmaxp[0,1]V1(L1(p,0.5))={p}\operatorname*{arg\,max}_{p\in[0,1]}V_{1}(L_{1}(p,0.5))=\{p^{\prime}\}, where p=0.707p^{\prime}=0.707 (see figure 9). Hence 0.5co¯(1(μ1(0.5)))0.5\notin\overline{co}(\mathscr{B}_{1}(\mu_{-1}(0.5))) and (0.5,0.5)mBBNE(0.5,0.5)\notin\mathrm{mBBNE}.

  3. (c)

    Let the CPT features for both the players be as in (b). Let the payoff matrix be as shown in figure 7, where β:=1/w1+(0.5)=1.414\beta:=1/w_{1}^{+}(0.5)=1.414 and γ=(1p)/p\gamma=(1-p^{\prime})/p^{\prime} (here p=0.707p^{\prime}=0.707 as in (b)). As observed in (b), 1(μ1(0.5))={p}\mathscr{B}_{1}(\mu_{-1}(0.5))=\{p^{\prime}\}. From the definition of γ\gamma, we see that player 22 is indifferent between her two actions, given her belief pp^{\prime} over player 11’s actions. Thus (p,0.5)(mNEBBNE)\pNE(p^{\prime},0.5)\in(\mathrm{mNE}\cap\mathrm{BBNE})\backslash\mathrm{pNE}.

  4. (d)

    Let ri=0,vi(x)=xr_{i}=0,v_{i}(x)=x, for i=1,2i=1,2. Let w1(p)=p0.5,w2+(p)=pw_{1}^{-}(p)=p^{0.5},w_{2}^{+}(p)=p. Let the payoff matrix be as shown in figure 7, where β:=1/w1(0.5)=1.414\beta:=1/w_{1}^{-}(0.5)=1.414. Note that the payoffs for player 11 are negations of her payoffs in (b), and her probability weighing function for losses is same as her probability weighing function for gains in (b). Thus her CPT value function V1(L1(p,q))V_{1}(L_{1}(p,q)) is the negation of her CPT value function in (b). In particular, we have V1(L1(0,0.5))=V1(L1(1,0.5))>V1(L1(p,0.5))V_{1}(L_{1}(0,0.5))=V_{1}(L_{1}(1,0.5))>V_{1}(L_{1}(p,0.5)) for all p(0,1)p\in(0,1). Thus, 0.5co¯(1(μ1(0.5)))0.5\in\overline{co}(\mathscr{B}_{1}(\mu_{-1}(0.5))), but 0.51(μ1(0.5))0.5\notin\mathscr{B}_{1}(\mu_{-1}(0.5)). The payoffs and CPT features of player 22 are same as in (b). Thus, (0.5,0.5)(mNEmBBNE)\BBNE(0.5,0.5)\in(\mathrm{mNE}\cap\mathrm{mBBNE})\backslash\mathrm{BBNE}.

  5. (e)

    Let the CPT features for both the players be as in (b). Let the payoff matrix be as shown in figure 7, where β:=1/w1+(0.5)=1.414,ϵ=0.1\beta:=1/w_{1}^{+}(0.5)=1.414,\epsilon=0.1, and γ:=(1p~)/p~\gamma:=(1-\tilde{p})/\tilde{p}; here p~=0.582\tilde{p}=0.582 is the unique maximizer of V1(L1(p,0.5))V_{1}(L_{1}(p,0.5)) (see figure 9). We have V1(L1(0,0.5))=2.071>2=V1(L1(1,0.5))V_{1}(L_{1}(0,0.5))=2.071>2=V_{1}(L_{1}(1,0.5)) and argmaxpV1(L1(p,0.5))={p~}\operatorname*{arg\,max}_{p}V_{1}(L_{1}(p,0.5))=\{\tilde{p}\} with V1(L1(p~,0.5))=2.125V_{1}(L_{1}(\tilde{p},0.5))=2.125. From the definition of γ\gamma, we see that player 22 is indifferent between her two actions, given her belief p~\tilde{p} over player 11’s actions. Thus, (p~,0.5)BBNE\mNE(\tilde{p},0.5)\in\mathrm{BBNE}\backslash\mathrm{mNE}.

  6. (f)

    Let the CPT features be as in example 3.9. Let p=0.996p^{*}=0.996 and q=0.340q^{*}=0.340 be the same as in example 3.9. Let the payoff matrix be as shown in figure 7. Note that the payoffs for both the players are the same as in example 3.9. Recall 1(μ1(q))=0\mathscr{B}_{1}(\mu_{-1}(q))=0 for q<qq<q^{*}, 1(μ1(q))={0,p}\mathscr{B}_{1}(\mu_{-1}(q))=\{0,p^{*}\} for q=qq=q^{*}, and 1(μ1(q))[p,1]\mathscr{B}_{1}(\mu_{-1}(q))\subset[p^{*},1] for q>qq>q^{*}, and hence 0.5co¯(1(μ1(q)))0.5\in\overline{co}(\mathscr{B}_{1}(\mu_{-1}(q^{*}))) and 0.51(μ1(q))0.5\notin\mathscr{B}_{1}(\mu_{-1}(q^{*})). Further, from the definition of γ\gamma, we have V2(L2(0.5,0))=V2(L2(0.5,q))=V2(L2(0.5,1)),q[0,1]V_{2}(L_{2}(0.5,0))=V_{2}(L_{2}(0.5,q))=V_{2}(L_{2}(0.5,1)),\forall q\in[0,1]. Hence, (0.5,q)mBBNE\(mNEBBNE)(0.5,q^{*})\in\mathrm{mBBNE}\backslash(\mathrm{mNE}\cap\mathrm{BBNE}).

  7. (g)

    Finally, if we let the players have EUT preferences and the payoffs as in (a), then (1,0)(mNEmBBNE)(1,0)\notin(\mathrm{mNE}\cup\mathrm{mBBNE}).

0 1
0 1,11,1 0,00,0
1 0,00,0 0,00,0
(a) pNE\mathrm{pNE}
0 1
0 2β,02\beta,0 0,10,1
1 β+1,1\beta+1,1 1,01,0
(b) mNE\mBBNE\mathrm{mNE}\backslash\mathrm{mBBNE}
0 1
0 2β,02\beta,0 0,10,1
1 β+1,γ\beta+1,\gamma 1,01,0
(c) (mNEBBNE)\pNE(\mathrm{mNE}\cap\mathrm{BBNE})\backslash\mathrm{pNE}
0 1
0 2β,0-2\beta,0 0,10,1
1 (β+1),1-(\beta+1),1 1,0-1,0
(d) (mNEmBBNE)\BBNE(\mathrm{mNE}\cap\mathrm{mBBNE})\backslash\mathrm{BBNE}
0 1
0 2β+ϵ,02\beta+\epsilon,0 0,10,1
1 β+1,γ\beta+1,\gamma 1,01,0
(e) BBNE\mNE\mathrm{BBNE}\backslash\mathrm{mNE}
0 1
0 4,04,0 0,10,1
1 3,13,1 1,01,0
(f) mBBNE\(mNEBBNE)\mathrm{mBBNE}\backslash(\mathrm{mNE}\cup\mathrm{BBNE})
Figure 7: Payoff matrices for the 2×22\times 2 games in example 3.15. The rows and the columns correspond to the actions of player 11 and player 22, respectively. In each cell, the left and right entries correspond to player 11 and player 22, respectively. The labels indicate the corresponding regions in figure 6. The game matrix for the example corresponding to region (g) is the same as that for the one corresponding to region (a).
00.20.20.40.40.60.60.80.811222.022.022.042.042.062.062.082.08ppV1(L1(p,0.5))V_{1}(L_{1}(p,0.5))
Figure 8: The CPT value function for player 11 in example 3.15(b), when q=0.5q=0.5 is the mixture of actions of player 22.
00.20.20.40.40.60.60.80.811222.052.052.12.1ppV1(L1(p,0.5))V_{1}(L_{1}(p,0.5))
Figure 9: The CPT value function for player 11 in example 3.15(e), when q=0.5q=0.5 is the mixture of actions of player 22.

4 Conclusion

In the study of non-cooperative game theory from a decision-theoretic viewpoint, it is important to distinguish between two types of randomization:

  1. 1.

    conscious randomizations implemented by the players, and

  2. 2.

    randomizations in conjectures resulting from the beliefs held by the other players about the behavior of a given player.

This difference becomes evident when the preferences of the players over lotteries do not satisfy betweenness, a weakened form of independence property. We considered nn-player normal form games where players have CPT preferences, an important example of preference relation that does not satisfy betweenness. This gives rise to four types of Nash equilibrium notions, depending on the different types of randomizations. We defined these different notions of equilibrium and discussed the question of their existence. The results are summarized in table 1.

Type of Nash equilibrium Strategies Conjectures Always exists
Pure Nash equilibrium Pure actions Exact conjectures No
Mixed action Nash equilibrium Pure actions Mixed conjectures Yes (Keskin, 2016)
Black-box strategy Nash equilibrium Black box strategies Exact conjectures No (Example 3.9)
Mixed black-box strategy Nash equilibrium Black box strategies Mixed conjectures Yes (Theorem 3.12)
Table 1: Different types of Nash equilibrium when players have CPT preferences.

Appendix A Joint continuity of the concave hull of a jointly continuous function

Let Δm1\Delta^{m-1} and Δn1\Delta^{n-1} be simplices of the corresponding dimensions with the usual topologies. Let f:Δm1×Δn1f:\Delta^{m-1}\times\Delta^{n-1}\to\mathbb{R} be a continuous function on Δm1×Δn1\Delta^{m-1}\times\Delta^{n-1} (with the product topology). Let 𝒫(Δm1)\mathscr{P}(\Delta^{m-1}) denote the space of all probability measures on Δm1\Delta^{m-1} with the topology of weak convergence. Let g:Δm1×Δn1g:\Delta^{m-1}\times\Delta^{n-1}\to\mathbb{R} be given by

g(x,y):=sup{𝔼Xpf(X,y)|p𝒫(Δm1),𝔼Xpid(X)=x}.g(x,y):=\sup\left\{\mathbb{E}_{X\sim p}f(X,y)\big{|}p\in\mathscr{P}(\Delta^{m-1}),\mathbb{E}_{X\sim p}\mathop{\rm id}(X)=x\right\}.

where id:Δm1Δm1\mathop{\rm id}:\Delta^{m-1}\to\Delta^{m-1} is the identity function id(x):=x,xΔm1\mathop{\rm id}(x):=x,\forall x\in\Delta^{m-1} and the expectation is over a random variable XX taking values in Δm1\Delta^{m-1} with the distribution pp.

Proposition A.1.

The function g(x,y)g(x,y) is continuous on Δm1×Δn1\Delta^{m-1}\times\Delta^{n-1}.

Proof.

We first prove that the function g(x,y)g(x,y) is upper semi-continuous. Let xtxx_{t}\to x and ytyy_{t}\to y. Let {g(xtn,ytn)}\{g(x_{t_{n}},y_{t_{n}})\} be a convergent subsequence of {g(xt,yt)}\{g(x_{t},y_{t})\} with limit LL. It is enough to show that the limit Lg(x,y)L\leq g(x,y). Since for all nn the set {p𝒫(Δm1),𝔼Xpid(X)=xtn}\{p\in\mathscr{P}(\Delta^{m-1}),\mathbb{E}_{X\sim p}\mathop{\rm id}(X)=x_{t_{n}}\} is compact, we know that there exists ptn𝒫(Δm1)p_{t_{n}}\in\mathscr{P}(\Delta^{m-1}), such that g(xtn,ytn)=𝔼Xptn[f(X,ytn)]g(x_{t_{n}},y_{t_{n}})=\mathbb{E}_{X\sim p_{t_{n}}}[f(X,y_{t_{n}})] and 𝔼Xptn[id(X)]=xtn\mathbb{E}_{X\sim p_{t_{n}}}[\mathop{\rm id}(X)]=x_{t_{n}}. The sequence {ptn}\{p_{t_{n}}\} has a convergent subsequence, say ptnkp¯p_{t_{n_{k}}}\to\bar{p} (because 𝒫(Δm1)\mathscr{P}(\Delta^{m-1}) is a compact space). Now, 𝔼Xp¯[id(X)]=limk𝔼Xptnk[id(X)]=limkxtnk=x\mathbb{E}_{X\sim\bar{p}}[\mathop{\rm id}(X)]=\lim_{k}\mathbb{E}_{X\sim p_{t_{n_{k}}}}[\mathop{\rm id}(X)]=\lim_{k}x_{t_{n_{k}}}=x. Further, 𝔼Xptnk[f(X,ytnk)]𝔼Xp¯[f(X,y)]\mathbb{E}_{X\sim p_{t_{n_{k}}}}[f(X,y_{t_{n_{k}}})]\to\mathbb{E}_{X\sim\bar{p}}[f(X,y)], since the product distributions ptnk×1{ytnk}p_{t_{n_{k}}}\times\textbf{1}\{y_{t_{n_{k}}}\}, for all kk, on Δm1×Δn1\Delta^{m-1}\times\Delta^{n-1}, converge weakly to the product distribution p¯×1{y}\bar{p}\times\textbf{1}\{y\}. Thus, L=𝔼Xp¯[f(X,y)]g(x,y)L=\mathbb{E}_{X\sim\bar{p}}[f(X,y)]\leq g(x,y) and the function g(x,y)g(x,y) is upper-semicontinuous.

We now prove that the function g(x,y)g(x,y) is lower semi-continuous. Let xtxx_{t}\to x and ytyy_{t}\to y. The simplex Δm1\Delta^{m-1} can be triangulated into finitely many other simplices, say T1,,TkT_{1},\dots,T_{k}, whose vertices are xx and some m1m-1 of the mm vertices of Δm1\Delta^{m-1}. Let (xtn)(x_{t_{n}}) be any subsequence such that all xtnTjx_{t_{n}}\in T_{j} for some simplex. It is enough to show that the lim inf\liminf of the sequence {g(xtn,ytn)}\{g(x_{t_{n}},y_{t_{n}})\} is greater than or equal to g(x,y)g(x,y). Let the other vertices of TjT_{j} be e1,,em1e_{1},\dots,e_{m-1}. Let ztn=(ztn1,,ztnl)z_{t_{n}}=(z_{t_{n}}^{1},\dots,z_{t_{n}}^{l}) be the barycentric coordinates of xtnx_{t_{n}} with respect to the simplex TjT_{j}, i.e.

xtn=(1ztn1ztnm1)x+ztn1e1++ztnm1em1.x_{t_{n}}=(1-z_{t_{n}}^{1}-\dots-z_{t_{n}}^{m-1})x+z_{t_{n}}^{1}e_{1}+\dots+z_{t_{n}}^{m-1}e_{m-1}.

The function g(x,y)g(x,y) is concave in xx for any fixed yy by construction. We have,

g(xtn,ytn)(1ztn1ztnm1)g(x,ytn)+ztn1g(e1,ytn)++ztnm1g(em1,ytn).g(x_{t_{n}},y_{t_{n}})\geq(1-z_{t_{n}}^{1}-\dots-z_{t_{n}}^{m-1})g(x,y_{t_{n}})+z_{t_{n}}^{1}g(e_{1},y_{t_{n}})+\dots+z_{t_{n}}^{m-1}g(e_{m-1},y_{t_{n}}).

Since ztn(0,,0)z_{t_{n}}\to(0,\dots,0) and g(e1,ytn),,g(em1,ytn)g(e_{1},y_{t_{n}}),\dots,g(e_{m-1},y_{t_{n}}) are all finite we get,

lim infg(xtn,ytn)lim infg(x,ytn).\liminf g(x_{t_{n}},y_{t_{n}})\geq\liminf g(x,y_{t_{n}}).

Let p~𝒫(Δm1)\tilde{p}\in\mathscr{P}(\Delta^{m-1}) be such that 𝔼Xp~[f(X,y)]=g(x,y)\mathbb{E}_{X\sim\tilde{p}}[f(X,y)]=g(x,y) and 𝔼Xp~[id(X)]=x\mathbb{E}_{X\sim\tilde{p}}[\mathop{\rm id}(X)]=x. Then, g(x,ytn)𝔼Xp¯[f(X,ytn)]g(x,y_{t_{n}})\geq\mathbb{E}_{X\sim\bar{p}}[f(X,y_{t_{n}})], for all nn, and hence,

lim infg(x,ytn)lim inf𝔼Xp~[f(X,ytn)]=g(x,y).\liminf g(x,y_{t_{n}})\geq\liminf\mathbb{E}_{X\sim\tilde{p}}[f(X,y_{t_{n}})]=g(x,y).

This shows that the function g(x,y)g(x,y) is lower semi-continuous.

Since the function g(x,y)g(x,y) is upper and lower semi-continuous, it is continuous. ∎

Appendix B An interesting functional equation

Lemma B.1.

Let w:[0,1][0,1]w:[0,1]\to[0,1] be a continuous, strictly increasing function such that w(0)=0w(0)=0 and w(1)=1w(1)=1. For any 0a1<c1<b<c2<a210\leq a_{1}<c_{1}<b<c_{2}<a_{2}\leq 1 such that (a2b)(bc1)=(ba1)(c2b)(a_{2}-b)(b-c_{1})=(b-a_{1})(c_{2}-b), let

[w(a2)w(b)]\displaystyle\left[w(a_{2})-w(b)\right] [w(b)w(c1)]=[w(b)w(a1)][w(c2)w(b)].\displaystyle\left[w(b)-w(c_{1})\right]=\left[w(b)-w(a_{1})\right]\left[w(c_{2})-w(b)\right]. (B.1)

Then w(p)=pw(p)=p for all p[0,1]p\in[0,1].

Proof.

Taking a1=0,c1=1/4,b=1/2,c2=3/4a_{1}=0,c_{1}=1/4,b=1/2,c_{2}=3/4 and a2=1a_{2}=1 in (B.1) we get,

[1w(1/2)]\displaystyle\left[1-w(1/2)\right] [w(1/2)w(1/4)]=[w(1/2)][w(3/4)w(1/2)],\displaystyle\left[w(1/2)-w(1/4)\right]=\left[w(1/2)\right]\left[w(3/4)-w(1/2)\right],

and hence,

w(3/4)=w(1/2)+w(1/2)w(1/4)w(1/4)w(1/2).w(3/4)=\frac{w(1/2)+w(1/2)w(1/4)-w(1/4)}{w(1/2)}.

Note that w(1/2)>0w(1/2)>0. Taking a1=0,c1=1/4,b=1/3,c2=1/2a_{1}=0,c_{1}=1/4,b=1/3,c_{2}=1/2 and a2=1a_{2}=1 in (B.1) we get,

[1w(1/3)]\displaystyle\left[1-w(1/3)\right] [w(1/3)w(1/4)]=[w(1/3)][w(1/2)w(1/3)],\displaystyle\left[w(1/3)-w(1/4)\right]=\left[w(1/3)\right]\left[w(1/2)-w(1/3)\right],

and hence,

w(1/3)=w(1/4)1w(1/2)+w(1/4).w(1/3)=\frac{w(1/4)}{1-w(1/2)+w(1/4)}.

Note that 1w(1/2)+w(1/4)>1w(1/2)>01-w(1/2)+w(1/4)>1-w(1/2)>0. Taking a1=0,c1=1/3,b=1/2,c2=2/3a_{1}=0,c_{1}=1/3,b=1/2,c_{2}=2/3 and a2=1a_{2}=1 in (B.1) we get,

[1w(1/2)]\displaystyle\left[1-w(1/2)\right] [w(1/2)w(1/3)]=[w(1/2)][w(2/3)w(1/2)],\displaystyle\left[w(1/2)-w(1/3)\right]=\left[w(1/2)\right]\left[w(2/3)-w(1/2)\right],

and substituting for w(1/3)w(1/3) we get,

w(2/3)=w(1/2)w(1/2)2+2w(1/2)w(1/4)w(1/4)w(1/2)w(1/2)2+w(1/2)w(1/4).w(2/3)=\frac{w(1/2)-w(1/2)^{2}+2w(1/2)w(1/4)-w(1/4)}{w(1/2)-w(1/2)^{2}+w(1/2)w(1/4)}.

Note that

w(1/2)w(1/2)2+w(1/2)w(1/4)=w(1/2)[1w(1/2)+w(1/4)]>0.w(1/2)-w(1/2)^{2}+w(1/2)w(1/4)=w(1/2)[1-w(1/2)+w(1/4)]>0.

Taking a1=0,c1=1/2,b=2/3,c2=3/4a_{1}=0,c_{1}=1/2,b=2/3,c_{2}=3/4 and a2=1a_{2}=1 in (B.1) we get,

[1w(2/3)]\displaystyle\left[1-w(2/3)\right] [w(2/3)w(1/2)]=[w(2/3)][w(3/4)w(2/3)].\displaystyle\left[w(2/3)-w(1/2)\right]=\left[w(2/3)\right]\left[w(3/4)-w(2/3)\right].

Simplifying we get,

w(2/3)w(2/3)w(3/4)=w(1/2)w(1/2)w(2/3),\displaystyle w(2/3)-w(2/3)w(3/4)=w(1/2)-w(1/2)w(2/3),

Substituting for w(2/3)w(2/3) and w(3/4)w(3/4) we get,

[w(1/2)w(1/2)2+2w(1/2)w(1/4)w(1/4)w(1/2)w(1/2)2+w(1/2)w(1/4)][w(1/4)w(1/2)w(1/4)w(1/2)]\displaystyle\left[\frac{w(1/2)-w(1/2)^{2}+2w(1/2)w(1/4)-w(1/4)}{w(1/2)-w(1/2)^{2}+w(1/2)w(1/4)}\right]\left[\frac{w(1/4)-w(1/2)w(1/4)}{w(1/2)}\right]
=w(1/2)[w(1/4)w(1/2)w(1/4)w(1/2)w(1/2)2+w(1/2)w(1/4)].\displaystyle=w(1/2)\left[\frac{w(1/4)-w(1/2)w(1/4)}{w(1/2)-w(1/2)^{2}+w(1/2)w(1/4)}\right].

Since w(1/4)w(1/2)w(1/4)>0w(1/4)-w(1/2)w(1/4)>0 and w(1/2)w(1/2)2+w(1/2)w(1/4)>0w(1/2)-w(1/2)^{2}+w(1/2)w(1/4)>0, we get

w(1/2)w(1/4)=2w(1/2)[w(1/2)w(1/4)].\displaystyle w(1/2)-w(1/4)=2w(1/2)[w(1/2)-w(1/4)].

Since w(1/2)w(1/4)>0w(1/2)-w(1/4)>0, we get w(1/2)=1/2w(1/2)=1/2.

For any fixed 0x<y10\leq x<y\leq 1, let

w(p):=w(p(yx)+x)w(x)w(y)w(x), for all 0p1.w^{\prime}(p^{\prime}):=\frac{w(p^{\prime}(y-x)+x)-w(x)}{w(y)-w(x)},\text{ for all }0\leq p^{\prime}\leq 1.

Note that w:[0,1][0,1]w^{\prime}:[0,1]\to[0,1] is a continuous, strictly increasing function with w(0)=0w^{\prime}(0)=0 and w(1)=1w^{\prime}(1)=1. Further, if 0a1<c1<b<c2<a210\leq a^{\prime}_{1}<c^{\prime}_{1}<b^{\prime}<c^{\prime}_{2}<a^{\prime}_{2}\leq 1 are such that (a2b)(bc1)=(ba1)(c2b)(a^{\prime}_{2}-b^{\prime})(b^{\prime}-c^{\prime}_{1})=(b^{\prime}-a^{\prime}_{1})(c^{\prime}_{2}-b^{\prime}), then

[w(a2)w(b)]\displaystyle\left[w^{\prime}(a^{\prime}_{2})-w^{\prime}(b^{\prime})\right] [w(b)w(c1)]=[w(b)w(a1)][w(c2)w(b)].\displaystyle\left[w^{\prime}(b^{\prime})-w^{\prime}(c^{\prime}_{1})\right]=\left[w^{\prime}(b^{\prime})-w^{\prime}(a^{\prime}_{1})\right]\left[w^{\prime}(c^{\prime}_{2})-w^{\prime}(b^{\prime})\right].

Thus w(1/2)=1/2w^{\prime}(1/2)=1/2 and hence w((x+y)/2)=(w(x)+w(y))/2.w\left((x+y)/{2}\right)=(w(x)+w(y))/{2}. Using this repeatedly we get w(k/2t)=k/2t,w({k}/{2^{t}})={k}/{2^{t}}, for 0k2t0\leq k\leq 2^{t}, t=1,2,t=1,2,\dots. Continuity of ww then implies w(p)=pw(p)=p, for all p[0,1]p\in[0,1]. ∎

References

  • Agranov and Ortoleva [2017] M. Agranov and P. Ortoleva. Stochastic choice and preferences for randomization. Journal of Political Economy, 125(1):40–68, 2017.
  • Allais [1953] M. Allais. L’extension des théories de l’équilibre économique général et du rendement social au cas du risque. Econometrica, Journal of the Econometric Society, pages 269–290, 1953.
  • Aumann and Brandenburger [1995] R. Aumann and A. Brandenburger. Epistemic conditions for Nash equilibrium. Econometrica: Journal of the Econometric Society, pages 1161–1180, 1995.
  • Bordley and Hazen [1991] R. Bordley and G. B. Hazen. SSB and weighted linear utility as expected utility with suspicion. Management Science, 37(4):396–408, 1991.
  • Bordley [1992] R. F. Bordley. An intransitive expectations-based Bayesian variant of prospect theory. Journal of Risk and Uncertainty, 5(2):127–144, 1992.
  • Camerer and Ho [1994] C. F. Camerer and T.-H. Ho. Violations of the betweenness axiom and nonlinearity in probability. Journal of risk and uncertainty, 8(2):167–196, 1994.
  • Chew and MacCrimmon [1979] S. Chew and K. MacCrimmon. Alpha-nu choice theory: an axiomatization of expected utility. University of British Columbia Faculty of Commerce working paper, 669, 1979.
  • Chew [1983] S. H. Chew. A generalization of the quasilinear mean with applications to the measurement of income inequality and decision theory resolving the Allais paradox. Econometrica: Journal of the Econometric Society, pages 1065–1092, 1983.
  • Chew [1989] S. H. Chew. Axiomatic utility theories with the betweenness property. Annals of operations Research, 19(1):273–298, 1989.
  • Dekel [1986] E. Dekel. An axiomatic characterization of preferences under uncertainty: Weakening the independence axiom. Journal of Economic theory, 40(2):304–318, 1986.
  • Dwenger et al. [2012] N. Dwenger, D. Kübler, and G. Weizsäcker. Flipping a coin: Theory and evidence. 2012.
  • Fishburn [1988] P. C. Fishburn. Nonlinear preference and utility theory. Number 5. Johns Hopkins University Press Baltimore, 1988.
  • Gul [1991] F. Gul. A theory of disappointment aversion. Econometrica: Journal of the Econometric Society, pages 667–686, 1991.
  • Kakutani [1941] S. Kakutani. A generalization of Brouwer’s fixed point theorem. Duke mathematical journal, 8(3):457–459, 1941.
  • Keskin [2016] K. Keskin. Equilibrium notions for agents with cumulative prospect theory preferences. Decision Analysis, 13(3):192–208, 2016.
  • Machina [1992] M. J. Machina. Choice under uncertainty: Problems solved and unsolved. In Foundations of Insurance Economics, pages 49–82. Springer, 1992.
  • Machina [2014] M. J. Machina. Nonexpected utility theory. Wiley StatsRef: Statistics Reference Online, 2014.
  • Nash [1951] J. Nash. Non-cooperative games. Annals of Mathematics, pages 286–295, 1951.
  • Nash [1950] J. F. Nash. Equilibrium points in n-person games. Proceedings of the national academy of sciences, 36(1):48–49, 1950.
  • Phade and Anantharam [2019] S. R. Phade and V. Anantharam. On the geometry of Nash and correlated equilibria with cumulative prospect theoretic preferences. Decision Analysis, 16(2):142–156, 2019.
  • Prelec [1990] D. Prelec. A “pseudo-endowment” effect, and its implications for some recent nonexpected utility models. Journal of Risk and Uncertainty, 3(3):247–259, 1990.
  • Prelec [1998] D. Prelec. The probability weighting function. Econometrica, pages 497–527, 1998.
  • Rudin [1991] W. Rudin. Functional analysis. 1991. Internat. Ser. Pure Appl. Math, 1991.
  • Sopher and Narramore [2000] B. Sopher and J. M. Narramore. Stochastic choice and consistency in decision making under risk: An experimental study. Theory and Decision, 48(4):323–350, 2000.
  • Tversky and Kahneman [1992] A. Tversky and D. Kahneman. Advances in prospect theory: Cumulative representation of uncertainty. Journal of Risk and uncertainty, 5(4):297–323, 1992.
  • Wakker [2010] P. P. Wakker. Prospect theory: For risk and ambiguity. Cambridge university press, 2010.
  • Weber and Camerer [1987] M. Weber and C. Camerer. Recent developments in modelling preferences under risk. Operations-Research-Spektrum, 9(3):129–151, 1987.