This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Optimizing stakes in simultaneous bets

Robbert Fokkink, Ludolf Meester, and Christos Pelekis
Abstract.

We want to find the convex combination SS of iid Bernoulli random variables that maximizes 𝐏(St)\mathbf{P}(S\geq t) for a given threshold tt. Endre Csóka conjectured that such an SS is an average if tpt\geq p, where pp is the success probability of the Bernoulli random variables. We prove this conjecture for a range of pp and tt.

Key words and phrases:
bold play, intersecting family, stochastic inequality, tail probability.
2010 Mathematics Subject Classification:
60G50, 05D05

We study tail probabilities of convex combinations of iid Bernoulli random variables. More specifically, let β1,β2,\beta_{1},\beta_{2},\ldots be an infinite sequence of independent Bernoulli random variables with success probability pp, and let tpt\geq p be a real number. We consider the problem of maximizing 𝐏(ciβit)\mathbf{P}(\sum c_{i}\beta_{i}\geq t) over all sequences c1,c2,c_{1},c_{2},\ldots of non-negative reals such that ci=1\sum c_{i}=1. By the weak law of large numbers, the supremum of 𝐏(ciβit)\mathbf{P}(\sum c_{i}\beta_{i}\geq t) is equal to 11 if t<pt<p. That is why we restrict our attention to tpt\geq p.

As a motivating example, consider a venture capitalist who has a certain fortune ff to invest in any number of startup companies. Each startup has an (independent) probability pp of succeeding, in which case it yields a return rr on investment. If the capitalist divides his fortune into a (possibly infinite!) sequence fif_{i} of investments, then his total return is rfiβi\sum rf_{i}\beta_{i}. Suppose he wants to maximize the probability that the total return reaches a threshold dd. Then we get our problem with t=drft=\frac{d}{rf}.

The problem has a how-to-gamble-if-you-must flavor [6]: the capitalist places stakes cic_{i} on a sequence of simultaneous bets. There is no need to place stakes higher than tt. The way to go all out, i.e., bold play, is to stake tt on 1t\lfloor\frac{1}{t}\rfloor bets, but this is not a convex combination. That is why we say that staking 1k\frac{1}{k} on kk bets with k=1tk={\lfloor\frac{1}{t}\rfloor} is bold play.

In a convex combination ciβi\sum c_{i}\beta_{i} we order c1c2c3c_{1}\geq c_{2}\geq c_{3}\geq\ldots. We denote the sequence (ci)(c_{i}) by γ\gamma and write Sγ=ciβiS_{\gamma}=\sum c_{i}\beta_{i}. We study the function

(1) π(p,t)=sup{𝐏(Sγt)γ}\pi(p,t)=\sup\left\{\mathbf{P}\left(S_{\gamma}\geq t\right)\mid\gamma\right\}

for 0pt10\leq p\leq t\leq 1. It is non-decreasing in pp and non-increasing in tt. The following has been conjectured by Csóka [5], who was inspired by some well known open problems in combinatorics:

Conjecture 1 (Csóka).

For every pp and tt there exists a kk\in\mathbb{N} such that π(p,t)\pi(p,t) is realized by ci=1kc_{i}=\frac{1}{k} if iki\leq k and ci=0c_{i}=0 if i>ki>k for some kk\in\mathbb{N}. In other words, the maximal probability is realized by an average.

If the conjecture is true, then π(p,t)\pi(p,t) is a binomial tail probability and we still need to determine the optimal kk. Numerical results of Csóka suggest that bold play is optimal for most parameter values.

We are able to settle the conjecture for certain parameter values, as illustrated in figure 1 below.

Refer to caption
Figure 1. The shaded region represents all (p,t)(p,t) for which we are able to settle the conjecture. In all these cases bold play is optimal. Our results can be divided into three parts: favorable odds p>12p>\frac{1}{2}, high threshold t12t\geq\frac{1}{2}, and unfavorable odds p<12p<\frac{1}{2}.

It is natural to expect, though we are unable to prove this, that a gambler becomes bolder if the threshold goes up or if the odds go down. In particular, if ppp^{\prime}\leq p and ttt^{\prime}\geq t and if bold play is optimal for (p,t)(p,t), then it is natural to expect that bold play is optimal for (p,t)(p^{\prime},t^{\prime}) as well. This is clearly visible in the figure above, which is a union of rectangles with lower right vertices (kk+1,kk+1)(\frac{k}{k+1},\frac{k}{k+1}) and (12k+1,1k+1)(\frac{1}{2k+1},\frac{1}{k+1}) for kk\in\mathbb{N}.

Our paper is organized as follows. We first lay the groundwork by analyzing properties of π(p,t)\pi(p,t) and prove that the supremum in equation 1 is a maximum. Then we cover the shaded region in figure 1 for the three separate parts of odds greater than one, threshold greater than half, and odds smaller than one. Finally, we recall an old result on binomial probabilities which would imply that (assuming Csóka’s conjecture holds and bold play is stable in the sense that we just explained) bold play is optimal if p1ntp\leq\frac{1}{n}\leq t for all nn\in\mathbb{N}.

1. Related problems and results

According to Csóka’s conjecture, if the coin is fixed and the stakes vary, then the maximum tail probability is attained by a (scaled) binomial. If the stake is fixed and the coins vary, then Chebyshev already showed that the maximum probability is attained by a binomial:

Theorem 2 (Chebyshev, [21]).

For a given ss and ll, let Z=β1++βlZ=\beta_{1}+\cdots+\beta_{l} be any sum of ll independent Bernoullis such that 𝐄[Z]=s\mathbf{E}[Z]=s. Then 𝐏(Zt)\mathbf{P}(Z\geq t) is maximized by Bernoullis for which the success probabilities assume at most three different values, only one of which is distinct from 0 and 11. In particular, the maximum 𝐏(Zt)\mathbf{P}(Z\geq t) is a binomial tail probability.

Samuels considered a more general situation with fixed expectations and arbitrary random variables.

Conjecture 3 (Samuels, [20]).

Let 0c1cl0\leq c_{1}\leq\cdots\leq c_{l} be such that i=1lci<1\sum_{i=1}^{l}c_{i}<1. Consider sup𝐏(X1++Xl1)\sup\mathbf{P}(X_{1}+\cdots+X_{l}\geq 1) over all collections of ll independent non-negative random variables such that 𝐄[Xi]=ci\mathbf{E}[X_{i}]=c_{i}. This supremum is a maximum which is attained by Xj=cjX_{j}=c_{j} for jkj\leq k and Xj=(1b)βjX_{j}=(1-b)\beta_{j} for j>kj>k, where kk is an integer, the βj\beta_{j} are Bernoulli random variables, and b=i=1kcib=\sum_{i=1}^{k}c_{i}. In other words, the gambler accumulates bb from small expectations before switching to bold play.

If all cic_{i} are equal, then the βj\beta_{j} are identically distributed, and the conjecture predicts that the maximum probability is attained by a binomial.

If one assumes that the Samuels conjecture holds, then one still needs to determine the optimal kk. If c1==cl=1l+1c_{1}=\ldots=c_{l}=\frac{1}{l+1} then the optimal kk is equal to zero  [1]. This implies that another well-known conjecture is a consequence of Samuels’ conjecture, see also [18].

Conjecture 4 (Feige, [7]).

For all collections of ll independent non-negative random variables such that 𝐄[Xi]1\mathbf{E}[X_{i}]\leq 1 it is true that

𝐏(X1++Xl<l+1)1e.\mathbf{P}(X_{1}+\cdots+X_{l}<l+1)\geq\frac{1}{e}.

As a step towards solving this conjecture, Feige proved the remarkable theorem that there exists a δ>0\delta>0 such that 𝐏(X1++Xl<l+1)δ\mathbf{P}(X_{1}+\cdots+X_{l}<l+1)\geq\delta. His original value of δ=113\delta=\frac{1}{13} has been gradually improved. The current best result is 0.17980.1798 by Guo et al [12].

2. Properties of π(p,t)\pi(p,t)

The function π(p,t)\pi(p,t) is defined on a region bounded by a rectangular triangle. It is easy to compute its value on the legs of the triangle: π(0,t)=0\pi(0,t)=0 and π(p,1)=p\pi(p,1)=p. It is much harder to compute the value on the hypothenuse.

Proposition 5.

12<π(p,p)<1\frac{1}{2}<\pi(p,p)<1 if 0<p<10<p<1.

Proof.

We follow the proof of [2, Lemma 1]. The following Paley-Zygmund type inequality for random variables of zero mean was proved in [13, lemma 2.2] and extended in [14]:

𝐏(X<0)(233)𝐄[X2]2𝐄[X4].\mathbf{P}(X<0)\geq\left(2\sqrt{3}-3\right)\frac{\mathbf{E}[X^{2}]^{2}}{\mathbf{E}[X^{4}]}.

Applying this to SγpS_{\gamma}-p we have

𝐏(Sγ<p)(233)𝐄[(Sγp)2]𝐄[(Sγp)4].\mathbf{P}(S_{\gamma}<p)\geq\left(2\sqrt{3}-3\right)\frac{\mathbf{E}[(S_{\gamma}-p)^{2}]}{\mathbf{E}[(S_{\gamma}-p)^{4}]}.

The second moment of SγpS_{\gamma}-p is equal to p(1p)ci2p(1-p)\sum c_{i}^{2} and the fourth moment is equal to

3p2(1p)2ijci2cj2+(p(1p)4+p4(1p))ci4\displaystyle 3p^{2}(1-p)^{2}\sum_{i\not=j}c_{i}^{2}c_{j}^{2}+(p(1-p)^{4}+p^{4}(1-p))\sum c_{i}^{4}

This can be bounded by

max(3,1p(1p)3)p2(1p)2(ci2)2\displaystyle\max\left(3,\frac{1}{p(1-p)}-3\right)p^{2}(1-p)^{2}\left(\sum c_{i}^{2}\right)^{2}

The Paley-Zygmund type inequality produces a lower bound on 𝐏(Sγ<p)\mathbf{P}(S_{\gamma}<p). Its complementary probability π(p,p)\pi(p,p) is bounded by:

π(p,p)1233max(3,1p(1p)3).\pi(p,p)\leq 1-\frac{2\sqrt{3}-3}{\max\left(3,\frac{1}{p(1-p)}-3\right)}.

It is possible to improve on this bound for small pp by using Feige’s theorem. We write Sγ=c1β1+SS_{\gamma}=c_{1}\beta_{1}+S. Then 𝐏(Sγ<p)𝐏(β1=0)𝐏(S<p)=(1p)𝐏(S<p)\mathbf{P}(S_{\gamma}<p)\geq\mathbf{P}(\beta_{1}=0)\mathbf{P}(S<p)=(1-p)\mathbf{P}(S<p). Note that 𝐄[S]=p(1c1)\mathbf{E}[S]=p(1-c_{1}) and we write 𝐏(S<p)=𝐏(S<𝐄[S]+pc1)\mathbf{P}(S<p)=\mathbf{P}(S<\mathbf{E}[S]+pc_{1}). We can approximate this probability by a truncated sum 𝐏(Sn<𝐄[S]+pc1)\mathbf{P}(S_{n}<\mathbf{E}[S]+pc_{1}), where Sn=c2β2++cnβnS_{n}=c_{2}\beta_{2}+\cdots+c_{n}\beta_{n} is a sum of independent random variables of expectation pc1\leq pc_{1}. By dividing by pc1pc_{1} and applying the bound 0.17980.1798 of [12] we find

π(p,p)0.8202+0.1798p.\pi(p,p)\leq 0.8202+0.1798p.

We have two upper bounds. The first is more restrictive for large pp and the second is more restrictive for small pp. The lower bound follows from bold play. Let kk\in\mathbb{N} be such that 1k+1<p1k\frac{1}{k+1}<p\leq\frac{1}{k}. If S¯k\bar{S}_{k} is the average of k1k\geq 1 Bernoullis, then 𝐏(S¯kp)=𝐏(S¯k1k)=1(1p)k>1(11k+1)k\mathbf{P}(\bar{S}_{k}\geq p)=\mathbf{P}(\bar{S}_{k}\geq\frac{1}{k})=1-(1-p)^{k}>1-(1-\frac{1}{k+1})^{k}. This is minimal and equal to 12\frac{1}{2} if k=1k=1. ∎

Refer to caption
Figure 2. The upper bounds by Paley-Zygmund’s inequality and Feige’s theorem, and the lower bound by bold play, as in the proof of Proposition 5. If Feige’s conjecture holds, then the upper bound would meet the lower bound at p=0p=0 in this figure. In corollary 16 we find π(12,12)=34\pi(\frac{1}{2},\frac{1}{2})=\frac{3}{4}, which is on the graph for the lower bound. If conjecture 1 is correct, then theorem 21 implies that bold play is optimal for p=t=1np=t=\frac{1}{n} for all nn. These are the tops of the zigzag.

We say that a sequence γ\gamma is finite if ci=0c_{i}=0 for all but finitely many ii, and infinite otherwise.

Proposition 6.

π(p,t)=sup{𝐏(Sγt)γisfinite}\pi(p,t)=\sup\left\{\mathbf{P}\left(S_{\gamma}\geq t\right)\mid\gamma\ \mathrm{is\ finite}\right\}

Proof.

According to Jessen and Wintner’s law of pure type [3, Theorem 3.5], either 𝐏(Sγ=s)=0\mathbf{P}(S_{\gamma}=s)=0 for each ss\in\mathbb{R} or there exists a countable set 𝒞\mathcal{C} such that 𝐏(Sγ𝒞)=1\mathbf{P}(S_{\gamma}\in\mathcal{C})=1. In other words, the random variable SγS_{\gamma} is either non-atomic or discrete. If XX and YY are independent, and if XX is non-atomic, then the convolution formula implies that X+YX+Y is non-atomic.

Suppose that γ\gamma is infinite. We prove that SγS_{\gamma} is non-atomic. Let (cij)(c_{i_{j}}) be a subsequence such that cij>2k=j+1cikc_{i_{j}}>2\sum_{k=j+1}^{\infty}c_{i_{k}}. Let II be the set of all iji_{j} and let JJ be its complement. Then both SI=IciβiS_{I}=\sum_{I}c_{i}\beta_{i} and SJ=JciβiS_{J}=\sum_{J}c_{i}\beta_{i} are either discrete or non-atomic. By our choice of cijc_{i_{j}}, SIS_{I} has the property that its value determines the values of all βi\beta_{i} for iIi\in I. This implies that SIS_{I} is non-atomic. Therefore, Sγ=SI+SJS_{\gamma}=S_{I}+S_{J} is non-atomic. In particular 𝐏(Sγt)=𝐏(Sγ>t)\mathbf{P}(S_{\gamma}\geq t)=\mathbf{P}(S_{\gamma}>t).

Denote a truncated sum by Sγ,n=inciβiS_{\gamma,n}=\sum_{i\leq n}c_{i}\beta_{i}. By monotonic convergence 𝐏(Sγ>t)=limn𝐏(Sγ,n>t)\mathbf{P}\left(S_{\gamma}>t\right)=\lim_{n\to\infty}\mathbf{P}(S_{\gamma,n}>t). Therefore, for any infinite γ\gamma, 𝐏(Sγt)\mathbf{P}(S_{\gamma}\geq t) can be approximated by tail probabilities of finite convex combinations. ∎

Csóka conjectures that the tail probability is maximized by an average. This would imply that the sup in proposition 6 is a max. We are unable to prove this, but we can prove that the sup in equation 1 is a max.

Theorem 7.

If p<tp<t then π(p,t)=𝐏(Sγt)\pi(p,t)=\mathbf{P}(S_{\gamma}\geq t) for some γ\gamma. Furthermore, π(p,t)\pi(p,t) is left-continuous in tt.

Proof.

We write π(p,t)=limstπ(p,s)\pi(p,t^{-})=\lim_{s\uparrow t}\pi(p,s). Since π(p,t)\pi(p,t) is decreasing in tt, it suffices to show that there exists an SγS_{\gamma} such that 𝐏(Sγt)π(p,t)\mathbf{P}(S_{\gamma}\geq t)\geq\pi(p,t^{-}). Let γn=(cn,i)i\gamma_{n}=(c_{n,i})_{i} be such that 𝐏(Sγntn)\mathbf{P}(S_{\gamma_{n}}\geq t_{n}) converges to π(p,t)\pi(p,t^{-}) for an increasing sequence tntt_{n}\uparrow t. By a standard diagonal argument we can assume that (cn,i)n(c_{n,i})_{n} is convergent for all ii. Let ci=limncn,ic_{i}=\lim_{n\to\infty}c_{n,i} and let γ=(ci)i\gamma=(c_{i})_{i}. Then γ\gamma is a non-increasing sequence which adds up to ci=1c1\sum c_{i}=1-c\leq 1. Observe that γ\gamma cannot be the all zero sequence, since this would imply that cn,10c_{n,1}\to 0 and Var(Sγn)=p(1p)cn,i2p(1p)cn,10\mathrm{Var}(S_{\gamma_{n}})=p(1-p)\sum c_{n,i}^{2}\leq p(1-p)c_{n,1}\to 0, so SγnS_{\gamma_{n}} converges to pp in distribution. Since we limit ourselves to p<tp<t, this means that 𝐏(Sγnt)0\mathbf{P}(S_{\gamma_{n}}\geq t)\to 0 which is nonsense. Therefore, 1c>01-c>0.

We first prove that π(p,t)𝐏(Sγtcp)\pi(p,t^{-})\leq\mathbf{P}(S_{\gamma}\geq t-cp). Fix an arbitrary ϵ>0\epsilon>0. Let i0i_{0} be such that ji0cj<ϵ4\sum_{j\geq i_{0}}c_{j}<\frac{\epsilon}{4} and ci0<ϵ4c_{i_{0}}<\epsilon^{4}. Let n0n_{0} be such that ji0|cn,jcj|<ϵ4\sum_{j\leq i_{0}}|c_{n,j}-c_{j}|<\frac{\epsilon}{4} and cn,i0<ϵ4c_{n,i_{0}}<\epsilon^{4} for all nn0n\geq n_{0}. Now

{Sγntn}{ji0cn,jβjtncpϵ}{ji0cn,jβjcp+ϵ}\textstyle\{S_{\gamma_{n}}\geq t_{n}\}\subset\left\{\sum_{j\leq i_{0}}c_{n,j}\beta_{j}\geq t_{n}-cp-\epsilon\right\}\bigcup\left\{\sum_{j\geq i_{0}}c_{n,j}\beta_{j}\geq cp+\epsilon\right\}

so by our assumptions

{Sγntn}{ji0cjβjtncp2ϵ}{ji0cn,jβjcp+ϵ}{Sγtncp2ϵ}{Tncp+ϵ}\textstyle\begin{array}[]{rcl}\{S_{\gamma_{n}}\geq t_{n}\}&\subset&\left\{\sum_{j\leq i_{0}}c_{j}\beta_{j}\geq t_{n}-cp-2\epsilon\right\}\bigcup\left\{\sum_{j\geq i_{0}}c_{n,j}\beta_{j}\geq cp+\epsilon\right\}\\ \\ &\subset&\left\{S_{\gamma}\geq t_{n}-cp-2\epsilon\right\}\bigcup\left\{T_{n}\geq cp+\epsilon\right\}\end{array}

where we write Tn=ji0cn,jβjT_{n}=\sum_{j\geq i_{0}}c_{n,j}\beta_{j}. Observe that

𝐄[Tn]=𝐄[Sγn]pj<i0cn,j<pp(j<i0cjϵ4)<pp(1cϵ2)<pc+ϵ2\textstyle\mathbf{E}[T_{n}]=\mathbf{E}[S_{\gamma_{n}}]-p\sum_{j<i_{0}}c_{n,j}<p-p\left(\sum_{j<i_{0}}c_{j}-\frac{\epsilon}{4}\right)<p-p\left(1-c-\frac{\epsilon}{2}\right)<pc+\frac{\epsilon}{2}

and

Var(Tn)=p(1p)ji0cn,j2cn,i0ji0cn,j<ϵ4.\mathrm{Var}(T_{n})=p(1-p)\sum_{j\geq i_{0}}c_{n,j}^{2}\leq c_{n,i_{0}}\sum_{j\geq i_{0}}c_{n,j}<\epsilon^{4}.

By Chebyshev’s inequality, we conclude that 𝐏(Tncp+ϵ)<ϵ\mathbf{P}\left(T_{n}\geq cp+\epsilon\right)<\epsilon for sufficiently small ϵ\epsilon. It follows that

𝐏(Sγntn)𝐏(Sγtncp2ϵ)+ϵ.\mathbf{P}(S_{\gamma_{n}}\geq t_{n})\leq\mathbf{P}\left(S_{\gamma}\geq t_{n}-cp-2\epsilon\right)+\epsilon.

By taking limits nn\to\infty and ϵ0\epsilon\to 0 we conclude that

π(p,t)𝐏(Sγtcp).\pi(p,t^{-})\leq\mathbf{P}\left(S_{\gamma}\geq t-cp\right).

Let γ¯=11cγ\bar{\gamma}=\frac{1}{1-c}\gamma. Then Sγ¯=11cSγS_{\bar{\gamma}}=\frac{1}{1-c}S_{\gamma} is a convex combination such that

𝐏(Sγ¯t)=𝐏(Sγ(1c)t)𝐏(Sγtcp)π(p,t).\mathbf{P}\left(S_{\bar{\gamma}}\geq t\right)=\mathbf{P}\left(S_{\gamma}\geq(1-c)t\right)\geq\mathbf{P}\left(S_{\gamma}\geq t-cp\right)\geq\pi(p,t^{-}).

Therefore, π(p,t)=𝐏(Sγ¯t)\pi(p,t)=\mathbf{P}(S_{\bar{\gamma}}\geq t) and these inequalities are equalities. ∎

We now more or less repeat this proof to show that π(p,t)\pi(p,t) is continuous in pp. Since we need to vary pp, we write βp\beta^{p} for a Bernoulli with succes probability pp and Sγp=ciβipS_{\gamma}^{p}=\sum c_{i}\beta_{i}^{p}.

Theorem 8.

π(p,t)\pi(p,t) is continuous in pp.

Proof.

For any ϵ>0\epsilon>0 choose a finite γ\gamma such that 𝐏(Sγpt)π(p,t)ϵ\mathbf{P}(S_{\gamma}^{p}\geq t)\geq\pi(p,t)-\epsilon. If pnp_{n} converges to pp then βpn\beta^{p_{n}} converges to βp\beta^{p} in probability. Since γ\gamma is finite

lim supnπ(pn,t)limn𝐏(Sγpnt)=𝐏(Sγpt)π(p,t)ϵ.\limsup_{n\to\infty}\pi(p_{n},t)\geq\lim_{n\to\infty}\mathbf{P}(S_{\gamma}^{p_{n}}\geq t)=\mathbf{P}(S_{\gamma}^{p}\geq t)\geq\pi(p,t)-\epsilon.

It follows that lim supnπ(pn,t)π(p,t)\limsup_{n\to\infty}\pi(p_{n},t)\geq\pi(p,t) for any sequence pnpp_{n}\to p. Since π(p,t)\pi(p,t) is increasing in pp, it follows that π(p,t)\pi(p,t) is left-continuous in pp.

We need to prove right continuity, i.e., π(p+,t)=π(p,t)\pi(p^{+},t)=\pi(p,t). This is trivially true on the hypothenuse, because this is the right-hand boundary of the domain. Consider p<tp<t. Let pnpp_{n}\downarrow p and γn\gamma_{n} be such that limn𝐏(Sγnpnt)=π(p+,t)\lim_{n\to\infty}\mathbf{P}(S_{\gamma_{n}}^{p_{n}}\geq t)=\pi(p^{+},t). By a standard diagonal argument we can assume that γn\gamma_{n} converges coordinatewise to some γ\gamma, which may not sum up to one. It cannot be the all zero sequence, i.e., the sum is not zero, by the same argument as in the proof of theorem 7. The sequence γ\gamma therefore sums up to 1c1-c for some 0c<10\leq c<1. Again, we split Sγp=H+TS^{p}_{\gamma}=H+T where H=ji0cjβjpH=\sum_{j\leq i_{0}}c_{j}\beta_{j}^{p} and T=j>i0cjβjpT=\sum_{j>i_{0}}c_{j}\beta_{j}^{p}. We choose i0i_{0} such that 𝐄[T]<ϵ4\mathbf{E}[T]<\frac{\epsilon}{4} and ci0<ϵ4c_{i_{0}}<\epsilon^{4}. Similarly, Sγn=Hn+TnS_{\gamma_{n}}=H_{n}+T_{n} where HnH_{n} converges to HH in probability, 𝐄[Tn]<pc+ϵ2\mathbf{E}[T_{n}]<pc+\frac{\epsilon}{2} and Var(Tn)<ϵ4\mathrm{Var}(T_{n})<\epsilon^{4} for sufficiently large nn. As in the previous proof, Chebyshev’s inequality and convergence in probability imply

𝐏(Sγnpnt)𝐏(Hntcpϵ)+ϵ𝐏(Htcp2ϵ)+ϵ.\mathbf{P}(S_{\gamma_{n}}^{p_{n}}\geq t)\leq\mathbf{P}(H_{n}\geq t-cp-\epsilon)+\epsilon\leq\mathbf{P}(H\geq t-cp-2\epsilon)+\epsilon.

for sufficiently large nn. By taking limits nn\to\infty and ϵ0\epsilon\to 0 it follows that π(p+,t)𝐏(Sγptcp)\pi(p^{+},t)\leq\mathbf{P}(S_{\gamma}^{p}\geq t-cp). If we standardize γ\gamma to a sequence γ¯\bar{\gamma} so that we get a convex combination, we again find that π(p+,t)𝐏(Sγ¯pt)\pi(p^{+},t)\leq\mathbf{P}(S_{\bar{\gamma}}^{p}\geq t). ∎

3. Favorable odds

We consider 12p<t\frac{1}{2}\leq p<t. In this case, bold play comes down to a single stake c1=1c_{1}=1. We say that I/nI\subset\mathbb{Z}/n\mathbb{Z} is an interval of length a<na<n if I=[b,b+a)={b,b+1,,b+a1}I=[b,b+a)=\{b,b+1,\ldots,b+a-1\} for some bb, which we call the initial element. We say that two intervals II and JJ are separate if IJI\cup J is not an interval. If \mathcal{F} is a family of sets, then we write \bigcup\mathcal{F} for the union of all these sets.

Lemma 9.

Let \mathcal{F} be a family of kk intervals of length aa in /n\mathbb{Z}/n\mathbb{Z} such that \bigcup\mathcal{F} is a proper subset. Then ||k+a1|\bigcup{\mathcal{F}}|\geq k+a-1.

Proof.

\bigcup\mathcal{F} is a union of say c1c\geq 1 separate intervals, all of lengths a\geq a. Any interval of length bab\geq a contains b(a1)b-(a-1) intervals of length aa. Therefore, \bigcup\mathcal{F} contains ||c(a1)|\bigcup\mathcal{F}|-c(a-1) intervals of length aa. It follows that ||c(a1)k|\bigcup\mathcal{F}|-c(a-1)\geq k. ∎

Two families \mathcal{F} and 𝒢\mathcal{G} are cross-intersecting if IJI\cap J\not=\emptyset for all II\in\mathcal{F} and J𝒢J\in\mathcal{G}.

Lemma 10.

Let \mathcal{F} be a family of kk intervals of length kk in /n\mathbb{Z}/n\mathbb{Z}. Let 𝒢\mathcal{G} be a family of intervals of length anka\leq n-k such \mathcal{F} and 𝒢\mathcal{G} are cross-intersecting. Then |𝒢|a|\mathcal{G}|\leq a.

Proof.

Let I=[b,b+k)I=[b,b+k) be any element in \mathcal{F}. An interval [c,c+a)[c,c+a) intersects II if and only if c[ba+1,b+k)c\in[b-a+1,b+k), which is an interval of length k+a1k+a-1. Therefore, the set \mathcal{I} of initial elements cc of intervals in 𝒢\mathcal{G} is contained in an intersection of kk intervals of length k+a1k+a-1. The complement of \mathcal{I} thus contains a union of kk intervals of length nka+1n-k-a+1. By the previous lemma, this union has cardinality na\geq n-a. Therefore, \mathcal{I} contains at most aa elements. ∎

Lemma 11.

Let (V,μ)(V,\mu) be a finite measure space such that μ(V)=b\mu(V)=b and let ViVV_{i}\subset V for i=1,,ki=1,\ldots,k be such that μ(Vi)t\mu(V_{i})\geq t. Then μ(Vi)kt(k1)b\mu\left(\bigcap V_{i}\right)\geq kt-(k-1)b.

Proof.
μ(Vi)=bμ(Vic)b(bμ(Vi))kt(k1)b.\mu\left(\bigcap V_{i}\right)=b-\mu\left(\bigcup V_{i}^{c}\right)\geq b-\sum(b-\mu(V_{i}))\geq kt-(k-1)b.

Theorem 12.

If kk+1<pk+1k+2<t\frac{k}{k+1}<p\leq\frac{k+1}{k+2}<t for some positive integer kk, then bold play is optimal.

Proof.

Bold play gives a probability pp of reaching tt. We need to prove that 𝐏(Sγt)p\mathbf{P}(S_{\gamma}\geq t)\leq p for arbitrary γ\gamma. By proposition 6 we may assume that γ\gamma is finite. It suffices to prove that 𝐏(Sγt)p\mathbf{P}(S_{\gamma}\geq t)\leq p for rational pp, since π(p,t)\pi(p,t) is monotonic in pp.

Let nn be the number of non-zero cic_{i} in γ\gamma and let p=abp=\frac{a}{b}. Let XiX_{i} be a sequence of nn independent discrete uniform U{0,b1}U\{0,b-1\} random variables, i.e, Xi=cX_{i}=c for c{0,,b1}c\in\{0,\ldots,b-1\} with probability 1b\frac{1}{b}. Let Bi0=1[0,a)(Xi)B_{i}^{0}=1_{[0,a)}(X_{i}) for 1in1\leq i\leq n. Then SγS_{\gamma} and Y0=ciBi0Y^{0}=\sum c_{i}B_{i}^{0} are identically distributed. Think of ciBi0c_{i}B_{i}^{0} as an assignment of weight cic_{i} to a random element in {0,,b1}=/b\{0,\ldots,b-1\}=\mathbb{Z}/b\mathbb{Z}. Let (j)\ell(j) be the sum of the coefficients – the load – that is assigned to j/bj\in\mathbb{Z}/b\mathbb{Z}. Then Y0=(0)++(a1)Y^{0}=\ell(0)+\cdots+\ell(a-1), i.e., Y0Y^{0} is the load of [0,a)[0,a). Instead of [0,a)[0,a) we might as well select any interval [j,j+a)/b[j,j+a)\subset\mathbb{Z}/b\mathbb{Z}. If YjY^{j} is the load of [j,j+a)[j,j+a), then SγYjS_{\gamma}\sim Y^{j}, and 𝐏(Sγt)=1b𝐏(Yjt)\mathbf{P}(S_{\gamma}\geq t)=\frac{1}{b}\sum\mathbf{P}(Y^{j}\geq t). We need to prove that 𝐏(Yjt)a\sum\mathbf{P}(Y^{j}\geq t)\leq a.

Let Ω\Omega be the sample space of the XiX_{i}. For ωΩ\omega\in\Omega, let J(ω)J(\omega) be the cardinality of 𝒥(ω)={j:Yj(ω)t}/b\mathcal{J}(\omega)=\{j\colon Y^{j}(\omega)\geq t\}\subset\mathbb{Z}/b\mathbb{Z}. In particular, 𝐏(Sγt)=1b𝐄[J]\mathbf{P}(S_{\gamma}\geq t)=\frac{1}{b}\mathbf{E}[J]. It suffices to prove that JaJ\leq a. Assume that J(ω)aJ(\omega)\geq a for some ωΩ\omega\in\Omega. Apply lemma 11 to the counting measure to find

|l=0k(𝒥(ω)la)|(k+1)akb.\left|\bigcap_{l=0}^{k}(\mathcal{J}(\omega)-la)\right|\geq(k+1)a-kb.

Note that i𝒥(ω)ji\in\mathcal{J}(\omega)-j if and only if [i+j,i+j+a)[i+j,i+j+a) has load t\geq t. Therefore, there are at least (k+1)akb(k+1)a-kb elements ii such that the intervals [i,i+a),[i+a,i+2a),,[i+ka,i+(k+1)a)[i,i+a),[i+a,i+2a),\ldots,[i+ka,i+(k+1)a) all have load t\geq t. The intersection of these k+1k+1 intervals is equal to

Ii=[i,i+(k+1)akb)I_{i}=[i,i+(k+1)a-kb)

It has load (k+1)tk\geq(k+1)t-k by lemma 11. Its complement IicI_{i}^{c} has load k+1(k+1)t<t\leq k+1-(k+1)t<t. If j𝒥(ω)j\in\mathcal{J}(\omega) then [j,j+a)[j,j+a) has load t\geq t and therefore it intersects IiI_{i} for all il=0k(J(ω)la)i\in\bigcap_{l=0}^{k}(J(\omega)-la). There are (k+1)akb\geq(k+1)a-kb such intervals IiI_{i}, and we denote this family by \mathcal{F}. Let 𝒢\mathcal{G} be the family of [j,j+a)[j,j+a) with jJ(ω)j\in J(\omega). Lemma 10 applies since the length of IiI_{i} is (k+1)akb(k+1)a-kb and since ab((k+1)akb)a\leq b-\left((k+1)a-kb\right). We conclude that J(ω)aJ(\omega)\leq a. ∎

With some additional effort, we can push this result to the hypothenuse.

Proposition 13.

If p=t=k+1k+2p=t=\frac{k+1}{k+2} for some positive integer kk, then bold play is optimal if k>1k>1, and c1=c2=c3=13c_{1}=c_{2}=c_{3}=\frac{1}{3} is optimal if k=1k=1.

Proof.

By proposition 6 it suffices to prove that 𝐏(Sγt)p\mathbf{P}(S_{\gamma}\geq t)\leq p for finite γ\gamma. We adopt the notation of the previous theorem. Let nn be the number of non-zero coefficients in γ\gamma, and let XiX_{i} be nn random selections of {0,,k+1}\{0,\ldots,k+1\}. We assign the coefficients according to these selections and let Yj=1(j)Y^{j}=1-\ell(j) be the load of the set {0,,k+1}{j}\{0,\ldots,k+1\}\setminus\{j\}. Each YjY^{j} is identically distributed to SγS_{\gamma}. For ωΩ\omega\in\Omega let J(ω)J(\omega) be the number of Yj(ω)Y^{j}(\omega) that reach the threshold, or equivalently, the number of loads (j)1k+2\ell(j)\leq\frac{1}{k+2}. We have 1k+2𝐄[J]=𝐏(Sγt)\frac{1}{k+2}\mathbf{E}[J]=\mathbf{P}(S_{\gamma}\geq t). In the proof above, we showed that Jk+1J\leq k+1 if t>p=k+1k+2t>p=\frac{k+1}{k+2}. This is no longer true now that we have t=pt=p. It may happen that J(ω)=k+2J(\omega)=k+2 in which case all Yj(ω)Y^{j}(\omega) are equal to k+1k+2\frac{k+1}{k+2} and all loads (j)\ell(j) are equal to 1k+2\frac{1}{k+2}. Note that this can only happen if all cic_{i} are bounded by 1k+2\frac{1}{k+2}, so nk+2n\geq k+2.

We think of the coefficients as being assigned one by one in increasing order. In particular, cn1c_{n-1} and cnc_{n} are placed last. If J=k+2J=k+2, then either kk or k+1k+1 of the loads are equal to 1k+2\frac{1}{k+2} before cn1c_{n-1} and cnc_{n} are placed. In the first case, there are two remaining loads <1k+2<\frac{1}{k+2} and the probability that cn1c_{n-1} are cnc_{n} are placed here is 2(k+2)2\frac{2}{(k+2)^{2}}. In the second case, there is only one remaining load <1k+2<\frac{1}{k+2} and the probability that cn1c_{n-1} and cnc_{n} are placed here is 1(k+2)2\frac{1}{(k+2)^{2}}. We conclude that 𝐏(J=k+2)2(k+2)2\mathbf{P}(J=k+2)\leq\frac{2}{(k+2)^{2}} and therefore

𝐄[J](k+2)𝐏(J=k+2)+(k+1)𝐏(J<k+2)=k+1+𝐏(J=k+2)\mathbf{E}[J]\leq(k+2)\mathbf{P}(J=k+2)+(k+1)\mathbf{P}(J<k+2)=k+1+\mathbf{P}(J=k+2)

is bounded by k+1+2(k+2)2k+1+\frac{2}{(k+2)^{2}}. Thus we obtain 𝐏(Sγt)k+1k+2+2(k+2)3\mathbf{P}(S_{\gamma}\geq t)\leq\frac{k+1}{k+2}+\frac{2}{(k+2)^{3}}. This bound is reached if k=1k=1 and c1=c2=c3=13c_{1}=c_{2}=c_{3}=\frac{1}{3}.

Let k>1k>1 and let J(ω)=k+2J(\omega)=k+2. We first consider the case that cn1cnc_{n-1}\not=c_{n}. suppose there are two remaining loads <1k+2<\frac{1}{k+2} before cn1c_{n-1} and cn2c_{n-2} are placed, then each cn1c_{n-1} and cnc_{n} can only be assigned to a unique place to complete all loads to 1k+2\frac{1}{k+2}. Since k>1k>1, there are at least two loads (i1)=(i2)=1k+2\ell(i_{1})=\ell(i_{2})=\frac{1}{k+2} before the final two coefficients are placed. Let ω¯\bar{\omega} assign cjc_{j} for j<n1j<n-1 in the same way as ω\omega, but it reassigns cn1c_{n-1} to i1i_{1} and cnc_{n} to i2i_{2}. Then J(ω¯)=kJ(\bar{\omega})=k, because the loads at i1i_{1} and i2i_{2} exceed the threshold. We can reconstruct ω\omega from ω¯\bar{\omega} because the loads at i1i_{1} and i2i_{2} are the only ones that exceed the threshold for ω¯\bar{\omega}, and their values are different because cn1cnc_{n-1}\not=c_{n}. We have a 111-1 correspondence between ω{J=k+2}\omega\in\{J=k+2\} and ω¯{J=k}\bar{\omega}\in\{J=k\}. Let ={J=k+2}\mathcal{E}=\{J=k+2\} and let ={ω¯:ω}\mathcal{F}=\{\bar{\omega}\colon\omega\in\mathcal{E}\}. Then 𝐏()=𝐏()\mathbf{P}(\mathcal{F})=\mathbf{P}(\mathcal{E}) and =\mathcal{E}\cap\mathcal{F}=\emptyset. This implies that

𝐄[J](k+2)𝐏()+k𝐏()+(k+1)𝐏(cc)k+1.\mathbf{E}[J]\leq(k+2)\mathbf{P}(\mathcal{E})+k\mathbf{P}(\mathcal{F})+(k+1)\mathbf{P}(\mathcal{E}^{c}\cap\mathcal{F}^{c})\leq k+1.

In particular 𝐏(Sγt)k+1k+2\mathbf{P}(S_{\gamma}\geq t)\leq\frac{k+1}{k+2} and bold play is optimal.

Finally, consider the remaining case k>1k>1 and J(ω)=k+2J(\omega)=k+2 and cn1=cnc_{n-1}=c_{n}. In this case, we may switch the assignments of cn1c_{n-1} and cnc_{n} to complete the loads. Let ωΩ\omega^{\prime}\in\Omega represent this switch (it may be equal to ω\omega if the assignments are the same). Again, let i1i_{1} and i2i_{2} be two locations for which the loads have already been completed before cn1c_{n-1} and cnc_{n} are placed. Let {ω¯,ω¯}\{\bar{\omega},\bar{\omega}^{\prime}\} be the elements which assign the first n2n-2 coefficients in the same way, but assigns the final two elements to i1i_{1} and i2i_{2}. In particular, J(ω¯)=J(ω¯)=kJ(\bar{\omega})=J(\bar{\omega}^{\prime})=k. We can reconstruct {ω,ω}\{\omega,\omega^{\prime}\} from {ω¯,ω¯}\{\bar{\omega},\bar{\omega}^{\prime}\}, the correspondence is injective, so again 𝐏()=𝐏()\mathbf{P}(\mathcal{E})=\mathbf{P}(\mathcal{F}) and we conclude in the same way that bold play is optimal. ∎

If bold play is stable, as discussed above, then proposition 13 would imply that bold play is optimal if pk+1k+2tp\leq\frac{k+1}{k+2}\leq t for k>1k>1. Establishing this stability of bold play appears to be hard, and we are able to settle only one specific case in the range p<k+1k+2=tp<\frac{k+1}{k+2}=t.

Proposition 14.

If p=2k+12k+3p=\frac{2k+1}{2k+3} and t=k+1k+2t=\frac{k+1}{k+2}, then bold play is optimal if k>1k>1.

Proof.

We randomly distribute the coefficients of a finite γ\gamma over 2k+32k+3 locations. Let Yj=(j)++(2k+j)Y^{j}=\ell(j)+\ldots+\ell(2k+j) be the load of the discrete interval [j,2k+j+1)[j,2k+j+1), where as before we reduce modulo 2k+32k+3. Then SγYjS_{\gamma}\sim Y^{j} and the sum of all YjY^{j} is equal to 2k+12k+1. Let JJ be the number of YjY^{j} that reach the threshold. Not all YjY^{j} can reach the threshold and therefore J2k+2J\leq 2k+2. Then

𝐏(Sγt)=𝐄[J]2k+32k+12k+3𝐏(J2k+1)+2k+22k+3𝐏(J=2k+2).\mathbf{P}(S_{\gamma}\geq t)=\frac{\mathbf{E}[J]}{2k+3}\leq\frac{2k+1}{2k+3}\mathbf{P}(J\leq 2k+1)+\frac{2k+2}{2k+3}\mathbf{P}(J=2k+2).

We need to prove that 𝐏(Sγt)2k+12k+3\mathbf{P}(S_{\gamma}\geq t)\leq\frac{2k+1}{2k+3}. If 𝐏(J=2k+2)=0\mathbf{P}(J=2k+2)=0 then we are done. Therefore, we may assume that 𝐏(J=2k+2)>0\mathbf{P}(J=2k+2)>0. Only one of the YjY^{j} does not meet the threshold and without loss of generality we may assume it is Y2Y^{2}, which has load 1(0)(1)1-\ell(0)-\ell(1). The other YjY^{j} reach the threshold, and since the sum of all YjY^{j} is equal to 2k+12k+1, we find that

2k+1(2k+2)t+1(0)(1).2k+1\geq(2k+2)t+1-\ell(0)-\ell(1).

In other words, (0)+(1)2k+2\ell(0)+\ell(1)\geq\frac{2}{k+2}. If (0)>1k+2\ell(0)>\frac{1}{k+2} then YjY^{j} does not reach the threshold if it does not include (0)\ell(0). Only 2k+12k+1 of the YjY^{j} include 0, contradicting our assumption that J=2k+2J=2k+2. Therefore (0)1k+2\ell(0)\leq\frac{1}{k+2} and since the same applies to (1)\ell(1) we have in fact that (0)=(1)=1k+2\ell(0)=\ell(1)=\frac{1}{k+2}. Since Yj=1(j1)(j2)Y^{j}=1-\ell(j-1)-\ell(j-2) we have that all (i)+(i+1)1k+2\ell(i)+\ell(i+1)\leq\frac{1}{k+2} other than (0)+(1)\ell(0)+\ell(1). In particular, (2)=(2k+2)=0\ell(2)=\ell(2k+2)=0. We find that Y2=kk+2Y^{2}=\frac{k}{k+2} and Y1=Y3=k+1k+2Y^{1}=Y^{3}=\frac{k+1}{k+2}. The sum of the remaining YjY^{j}, with j{1,2,3}j\not\in\{1,2,3\} is at least 2kt2kt and the sum of all YjY^{j} is 2k+12k+1. Since 2kt=2k+1kk+22k+22k+12kt=2k+1-\frac{k}{k+2}-\frac{2k+2}{2k+1} all the remaining YjY^{j} have to be equal to tt. It follows that the loads alternate between zero and 1k+2\frac{1}{k+2}: (i)=1k+2\ell(i)=\frac{1}{k+2} if ii is odd and (i)=0\ell(i)=0 if i>0i>0 is even. We conclude that if J=k+2J=k+2 then all non-zero loads are equal and only two non-zero loads are consecutive. There are exactly 2k+32k+3 such arrangements. There are also 2k+32k+3 arrangements in which the non-zero loads are consecutive. In this case J=k+22kJ=k+2\leq 2k. It follows that 𝐏(J2k)P(J=2k+2)\mathbf{P}(J\leq 2k)\geq P(J=2k+2), which implies that 𝐄[J]2k+1\mathbf{E}[J]\leq 2k+1. Bold play is optimal. ∎

These results conclude our analysis of the upper right hand block of figure 1. A zigzag of triangles along the hypothenuse remains. Numerical results of Csóka [5] suggest that bold play is optimal for all of these triangles, except for the one touching on {(p,p):12p23}\{(p,p)\colon\frac{1}{2}\leq p\leq\frac{2}{3}\}. In the next section we will confirm that bold play is not optimal for this triangle.

4. High threshold

We now consider the case p12<tp\leq\frac{1}{2}<t, when bold play comes down to a single stake c1=1c_{1}=1. We need to maximize 𝐏(Sγt)\mathbf{P}(S_{\gamma}\geq t) and we may assume that γ\gamma is finite by proposition 6. Suppose that γ\gamma has n\leq n non-zero coefficients, i.e., cn+1=0c_{n+1}=0. Let t,γ\mathcal{F}_{t,\gamma} be the family of V{1,2,,n}V\subset\{1,2,\ldots,n\}, such that iVcit\sum_{i\in V}c_{i}\geq t. Let p(V)=p|V|(1p)n|V|p(V)=p^{|V|}(1-p)^{n-|V|}. Then

(2) 𝐏(Sγt)=Vt,γp(V).\mathbf{P}(S_{\gamma}\geq t)=\sum_{V\in\mathcal{F}_{t,\gamma}}p(V).

Therefore we need to determine the family t,γ\mathcal{F}_{t,\gamma} that maximizes the sum on the right hand side. Problems of this type are studied in extremal combinatorics, see [8] for recent progress. A family \mathcal{F} is intersecting if no two elements are disjoint. Two standard examples of intersecting families are 1\mathcal{F}_{1}, the family of all VV such that 1V1\in V, and >n/2\mathcal{F}_{>n/2}, the family of all subsets such that |V|>n/2|V|>n/2. Fishburn et al [9] settled the problem of maximizing

p()=Vp(V)p(\mathcal{F})=\sum_{V\in\mathcal{F}}p(V)

over all intersecting families \mathcal{F}:

Theorem 15 (Fishburn et al).

For a fixed nn, let \mathcal{F} be any intersecting family of subsets from {1,,n}\{1,\ldots,n\}. If p12p\leq\frac{1}{2} then p()p(\mathcal{F}) is maximized by 1\mathcal{F}_{1}. If p12p\geq\frac{1}{2} and nn is odd, then p()p(\mathcal{F}) is maximized by >n/2\mathcal{F}_{>n/2}.

Proof.

Following [9]. First suppose nn is odd. At most one of VV and VcV^{c} can be in \mathcal{F}. If p12p\geq\frac{1}{2}, then p(V)>p(Vc)p(V)>p(V^{c}) if |V|>|Vc||V|>|V^{c}|. Therefore p()p(\mathcal{F}) is maximal if \mathcal{F} contains each set of largest cardinality. It follows that >n/2\mathcal{F}_{>n/2} maximizes p()p(\mathcal{F}) if nn is odd and p12p\geq\frac{1}{2}.

Now consider an arbitrary nn and p12p\leq\frac{1}{2}. Let ca=|a|c_{a}=|\mathcal{F}^{a}| be the cardinality of the subfamily a={V:|V|=a}.\mathcal{F}^{a}=\{V\in\mathcal{F}\colon|V|=a\}. At most one of VV and VcV^{c} can be in \mathcal{F} and therefore ca+cna(na)c_{a}+c_{n-a}\leq\binom{n}{a}. Since p12p\leq\frac{1}{2} we have p(V)>p(Vc)p(V)>p(V^{c}) if |V|<|Vc||V|<|V^{c}|. For a<n2a<\frac{n}{2} we want to maximize cac_{a} under the constraint ca+cna(na)c_{a}+c_{n-a}\leq\binom{n}{a} (for a=n2a=\frac{n}{2} it does not matter which of the two subsets we select, as long as we select one of them). By the Erdős-Ko-Rado theorem, if a<n2a<\frac{n}{2} then cac_{a} is maximized by a family of subsets that contain one common element. For such a family, ca+cna=(na)c_{a}+c_{n-a}=\binom{n}{a}. It follows that p(1)p(\mathcal{F}_{1}) is maximal if p12p\leq\frac{1}{2}. ∎

Note that 1\mathcal{F}_{1} corresponds to t,γ\mathcal{F}_{t,\gamma} if t>12t>\frac{1}{2} and c1=1c_{1}=1, i.e., bold play.

Corollary 16.

If p12<tp\leq\frac{1}{2}<t then bold play is optimal.

Proof.

If t>12t>\frac{1}{2} then t,γ\mathcal{F}_{t,\gamma} is intersecting. The maximizing family 1\mathcal{F}_{1} corresponds to γ\gamma with γ=(1,0,,0)\gamma=(1,0,\ldots,0). This takes care of the upper left-hand block in figure 1. ∎

The positive odds part of theorem 15 can be applied to the triangle touching on {(p,p):12p23}\{(p,p)\colon\frac{1}{2}\leq p\leq\frac{2}{3}\} that we mentioned above. The family >n/2\mathcal{F}_{>n/2} can be represented by t,γ\mathcal{F}_{t,\gamma} if n=2k+1n=2k+1 and 12<tk+12k+1\frac{1}{2}<t\leq\frac{k+1}{2k+1}, by taking γ=(12k+1,,12k+1)\gamma=(\frac{1}{2k+1},\ldots,\frac{1}{2k+1}).

Corollary 17.

If 12<pt23\frac{1}{2}<p\leq t\leq\frac{2}{3} then bold play is not optimal.

Proof.

Choose kk maximal such that tk+12k+1t\leq\frac{k+1}{2k+1} and set n=2k+1n=2k+1. Then >n/2\mathcal{F}_{>n/2} is the unique maximizer of p()p(\mathcal{F}) and corresponds to c1==c2k+1=12k+1c_{1}=\cdots=c_{2k+1}=\frac{1}{2k+1}, which is not bold play. ∎

Finally, we settle the remaining part of the box in the upper left corner of figure 1.

Corollary 18.

If pt=12p\leq t=\frac{1}{2} then bold play is optimal.

Proof.

Note that we may restrict our attention to γ=(c1,c2,)\gamma=(c_{1},c_{2},\ldots) such that c112c_{1}\leq\frac{1}{2} and that bold play corresponds to (12,12,0,)(\frac{1}{2},\frac{1}{2},0,\ldots).

𝐏(Sγ12)=p𝐏(Sγ12β1=1)+(1p)𝐏(Sγ12β1=0)p+(1p)𝐏(Sγ12β1=0)\begin{array}[]{ccl}\mathbf{P}(S_{\gamma}\geq\frac{1}{2})&=&p\mathbf{P}(S_{\gamma}\geq\frac{1}{2}\mid\beta_{1}=1)+(1-p)\mathbf{P}(S_{\gamma}\geq\frac{1}{2}\mid\beta_{1}=0)\\ &\leq&p+(1-p)\mathbf{P}(S_{\gamma}\geq\frac{1}{2}\mid\beta_{1}=0)\end{array}

If γ¯=11c1(c2,c3,)\bar{\gamma}=\frac{1}{1-c_{1}}(c_{2},c_{3},\ldots) then 𝐏(Sγ12β1=0)=𝐏(Sγ¯12(1c1))p\mathbf{P}(S_{\gamma}\geq\frac{1}{2}\mid\beta_{1}=0)=\mathbf{P}(S_{\bar{\gamma}}\geq\frac{1}{2(1-c_{1})})\leq p by theorem 15. We find that 𝐏(Sγ12)p+(1p)p\mathbf{P}(S_{\gamma}\geq\frac{1}{2})\leq p+(1-p)p with equality for bold play. ∎

These results take care of the box in the upper left corner of figure 1, including its boundaries.

5. Unfavorable odds

A family 2{1,,n}\mathcal{F}\subset 2^{\{1,\ldots,n\}} has matching number kk, denoted by ν()=k\nu(\mathcal{F})=k, if the maximum number of pairwise disjoint VV\in\mathcal{F} is equal to kk. In particular, \mathcal{F} is intersecting if and only if ν()=1\nu(\mathcal{F})=1. The matching number of t,γ\mathcal{F}_{t,\gamma} is bounded by 1t\frac{1}{t}, because jVcjt\sum_{j\in V}c_{j}\geq t for each Vt,γV\in\mathcal{F}_{t,\gamma} and γ\gamma sums up to one.

A family u\mathcal{F}^{u} is uu-uniform if all its elements have cardinality uu. According to the Erdős matching conjecture [1, 10, 11], if n(k+1)un\geq(k+1)u then the maximum cardinality |u||\mathcal{F}^{u}| of a uu-uniform family such that ν(u)k\nu(\mathcal{F}^{u})\leq k is either attained by ku\mathcal{F}_{k}^{u}, the family of all uu-subsets containing at least one element from {1,,k}\{1,\ldots,k\}, or by [(k+1)u1]u\mathcal{F}_{[(k+1)u-1]}^{u}, the family containing all uu-subsets from {1,,(k+1)u1}\{1,\ldots,(k+1)u-1\}. Frankl [10] proved that ku\mathcal{F}_{k}^{u} has maximum cardinality if n(2k+1)ukn\geq(2k+1)u-k. For recent progress on this conjecture, see [11] and the references therein.

Theorem 19.

If p<12k+1p<\frac{1}{2k+1} and 1k+1<t\frac{1}{k+1}<t for some positive integer kk then bold play is optimal.

Proof.

We need to prove that 𝐏(Sγt)1(1p)k\mathbf{P}(S_{\gamma}\geq t)\leq 1-(1-p)^{k} for finite γ=(c1,c2,)\gamma=(c_{1},c_{2},\ldots). For a large enough nn we have that cj=0c_{j}=0 if j>nj>n. We have

𝐏(Sγt)=t,γp(V)=j|t,γj|pj(1p)nj\mathbf{P}(S_{\gamma}\geq t)=\sum_{\mathcal{F}_{t,\gamma}}p(V)=\sum_{j}|\mathcal{F}_{t,\gamma}^{j}|p^{j}(1-p)^{n-j}

where |t,γj||\mathcal{F}_{t,\gamma}^{j}| denotes the number of subsets of cardinality jj. By Frankl’s result, we can put a bound on |t,γj||\mathcal{F}_{t,\gamma}^{j}| if (2k+1)jkn(2k+1)j-k\leq n. For larger jj we simply bound by (nj)\binom{n}{j}. In this way we get that 𝐏(Sγt)\mathbf{P}(S_{\gamma}\geq t) is bounded by

jn+k2k+1((nj)(nkj))pj(1p)nj+j>n+k2k+1(nj)pj(1p)nj\sum_{j\leq{\frac{n+k}{2k+1}}}\left(\binom{n}{j}-\binom{n-k}{j}\right)p^{j}(1-p)^{n-j}+\sum_{j>{\frac{n+k}{2k+1}}}\binom{n}{j}p^{j}(1-p)^{n-j}

which is equal to

1jn+k2k+1(nkj)pj(1p)nj=1(1p)k𝐏(Xn+k2k+1)1-\sum_{j\leq{\frac{n+k}{2k+1}}}\binom{n-k}{j}p^{j}(1-p)^{n-j}=1-(1-p)^{k}\mathbf{P}\left(X\leq\frac{n+k}{2k+1}\right)

for XBin(nk,p)X\sim\mathrm{Bin}(n-k,p). By our assumptions, there exists a c<1c<1 such that p<c2k+1p<\frac{c}{2k+1}. If nn\to\infty then 𝐏(Xn+k2k+1)1\mathbf{P}\left(X\leq\frac{n+k}{2k+1}\right)\to 1 since 𝐄[X]=(nk)p<(nk)c2k+1\mathbf{E}[X]=(n-k)p<\frac{(n-k)c}{2k+1}. ∎

We can push this result to the hypothenuse, using the same approach as in the proof of proposition 13, in one particular case: p=t=13p=t=\frac{1}{3}. By stability, one would expect that bold play is optimal for any p<t=13p<t=\frac{1}{3} but we can only prove this for p=1bp=\frac{1}{b} for integers b>3b>3.

Proposition 20.

Bold play is optimal if t=13t=\frac{1}{3} and p=1bp=\frac{1}{b} for an integer b3b\geq 3.

Proof.

We may assume that γ\gamma is finite and we randomly assign its coefficients to {0,1,,b1}\{0,1,\ldots,b-1\}. We denote (0)\ell(0), the load at zero, by YY and we need to prove that

𝐏(Y13)p+(1p)p+(1p)2p=r\mathbf{P}\left(Y\geq\frac{1}{3}\right)\leq p+(1-p)p+(1-p)^{2}p=r

which is the success probability of bold play if t=13t=\frac{1}{3}. Let KK be the number of loads exceeding the threshold of 13\frac{1}{3} before the last two coefficients cn1c_{n-1} and cnc_{n} are assigned. Obviously, KK is either equal to 0 or 11 or 22. We will show that 𝐏(Y13|K=j)r\mathbf{P}(Y\geq\frac{1}{3}|K=j)\leq r for j{0,1,2}j\in\{0,1,2\}.

Suppose that K=0K=0. In this case, if YY reaches the threshold, then at least one of the two final coefficients has to be placed in 0. This happens with probability p+(1p)p<rp+(1-p)p<r.

Suppose that K=1K=1. One load has reached the threshold before the final two coefficients are placed. This load is in 0 with probability pp. If the load is not in 0, then at least one of the remaining two coefficients has to be placed there. This happens with probability p+(1p)pp+(1-p)p. We conclude that

𝐏(Y13|K=1)p+(1p)(p+(1p)p)=r.\mathbf{P}\left(Y\geq\frac{1}{3}\,\middle|\,K=1\right)\leq p+(1-p)(p+(1-p)p)=r.

Suppose that K=2K=2. In other words, two loads have already reached the threshold before the final two coefficients are placed. The probability that one of these two loads is in 0 is 2p2p. If none of the two loads is in 0, then (0)\ell(0) can only reach the threshold if both remaining coefficients are assigned to 0. The probability that this happens is p2p^{2}.

𝐏(Y13|K=2)2p+(12p)p2r\mathbf{P}\left(Y\geq\frac{1}{3}\,\middle|\,K=2\right)\leq 2p+(1-2p)p^{2}\leq r

if p13p\leq\frac{1}{3}. ∎

6. Binomial tails

If conjecture 1 holds, then the tail probability is maximized by a Bernoulli average X¯k\bar{X}_{k} and we need to determine the optimal kk. It is more convenient to state this in terms of binomials. For a fixed pp and tt, maximize

𝐏(Bin(k,p)kt)\mathbf{P}\left(\mathrm{Bin}(k,p)\geq kt\right)

for a positive integer kk. Since the probability increases if kk increases and ktkt does not pass an integer, we may restrict our attention to kk such that ktn<(k+1)tkt\leq n<(k+1)t for some integer nn. In other words, we need to only consider k=ntk=\lfloor\frac{n}{t}\rfloor for nn\in\mathbb{N}. If t=1at=\frac{1}{a} is the reciprocal of an integer aa, then the kk are multiples of aa. This is a classical problem. In 1693 John Smith asked which kk is optimal if a=6a=6 and p=16p=\frac{1}{6}. Or in his original words, which of the following events is most likely: fling at least one six with 6 dice, or at least two sixes with 12 dice, or at least three sixes with 18 dice. The problem was communicated by Samuel Pepys to Isaac Newton, who computed the probabilities. Chaundy and Bullard [4] gave a very nice historical description (more history can be found in [16, 17]) and solved the problem, see also [15].

Theorem 21 (Chaundy and Bullard).

For an integer a>1a>1, 𝐏(Bin(ka,1a)k)\mathbf{P}(\mathrm{Bin}(ka,\frac{1}{a})\geq k) is maximal for k=1k=1. Even more so, the tail probabilities strictly decrease with kk.

In other words, if p=t=1ap=t=\frac{1}{a} and if Csóka’s conjecture holds, then bold play is optimal. By stability, one would expect that bold play is optimal for p1atp\leq\frac{1}{a}\leq t. It turns out that it is possible to extend Chaundy and Bullard’s theorem in this direction and prove that 𝐏(Bin(ka,p)k)\mathbf{P}(\mathrm{Bin}(ka,p)\geq k) decreases with kk for arbitrary p1ap\leq\frac{1}{a}, see [19, Theorem 1.5.4].

7. Conclusion

We settled Csóka’s conjecture for a range of parameters building on combinatorial methods. Csóka’s conjecture predicts that 𝐏(Sγt)\mathbf{P}(S_{\gamma}\geq t) attains its maximum at an extreme point of the subset of positive non-increasing sequences γ\gamma in the unit ball in 1\ell_{1}. Perhaps variational methods need to be considered.

Christos Pelekis was supported by the Czech Science Foundation (GAČR project 18-01472Y), by the Czech Academy of Sciences (RVO: 67985840), and by a visitor grant of the Dutch mathematics cluster Diamant.

References

  • [1] N. Alon, P. Frankl, H. Huang, V. Rödl, A. Ruciński, B. Sudakov, Large matchings in uniform hypergraphs and the conjectures of Erdős and Samuels. J. Combin. Th., Ser. A 119 (2012) 1200–1215.
  • [2] I. Arieli, Y. Babichenko, R. Peretz, H. Peyton Young, The speed of innovation in social networks. Econometrica 88 no. 2 (2020), 569–594.
  • [3] L. Breiman, Probability, Reading, Addison-Wesley, 1968.
  • [4] T.W. Chaundy, J.E. Bullard, John Smith’s problem. Math. Gazette 44(350), (1960), 253–260.
  • [5] E. Csóka, Limit theory of discrete mathematics problems. arXiv1505.06984.
  • [6] L. E. Dubins and L. J. Savage, Inequalities for Stochastic Processes (How to Gamble If You Must). Dover, 1976.
  • [7] U. Feige, On sums of independent random variables with unbounded variances, and estimating the average degree in a graph. SIAM J. Comput. 35 (2006), 964–984.
  • [8] Y. Filmus, The weighted complete intersection theorem. J. Combin. Th. Ser. A 151 (2017), 84–101.
  • [9] P.C. Fishburn, P. Frankl, D. Freed, J. Lagarias, A.M. Odlyzko. Probabilities for intersecting systems and random subsets of finite sets. SIAM. J. Algebraic Discrete Methods 7 no. 1 (1986) 73–79.
  • [10] P. Frankl, Improved bounds on Erdős’ matching conjecture. J. Combin. Th., ser. A, 120 (2013) 1068–1072.
  • [11] P. Frankl, A. Kupavskii, Some results around the Erdős matching conjecture. Acta Math. Univ. Comenianae 88 no. 3 (2019), 695–699.
  • [12] J. Guo, S. He, Z. Ling, Y. Liu, Bounding probability of small deviation on sum of independent random variables: Combination of moment approach and Berry-Esseen theorem. arXiv:2003.03197.
  • [13] S. He, Z. Q. Luo, J. Nie, S. Zhang, Semidefinite relaxation bounds for indefinite homogeneous quadratic optimization. SIAM J. Optim. 19 (2007) 503–523.
  • [14] S. He, J. Zhang, S. Zhang, Bounding probability of small deviation: a fourth moment approach. Math. Oper. Res. 45 no. 1 (2010), 208–232.
  • [15] K. Jogdeo, S.M. Samuels, Monotone Convergence of Binomial Probabilities and a Generalization of Ramanujan’s Equation. Ann. Math. Statist. 39 no. 4, (1968), 1191–1195.
  • [16] T.H. Koornwinder, M.J. Schlosser, On an identity by Chaundy and Bullard, I. Indag. Math. (NS) 19, (2008), 239–261.
  • [17] T.H. Koornwinder, M.J. Schlosser, On an identity by Chaundy and Bullard, II. Indag. Math. (NS) 24, (2013), 174–180.
  • [18] R. Paulin, On some conjectures by Feige and Samuels. arXiv1703.05152.
  • [19] C. Pelekis, Search games on hypergraphs. PhD thesis, TU Delft, 2014.
  • [20] S.M. Samuels, On a Chebyshev-type inequality for sums of independent random variables. Ann. Math. Statist. 37 no. 1, (1966), 248–259.
  • [21] P. Tchebichef, Démonstration élémentaire d’une proposition générale de la theorie des probabilités. J. Reine. Angew. Math. 33 (1848), 259–267.

Institute of Applied Mathematics
Delft University of Technology
Mourikbroekmanweg 6
2628 XE Delft, The Netherlands
r.j.fokkink@tudelft.nl, l.e.meester@tudelft.nl
School of Electrical and Computer Engineering
National Technical University of Athens
Zografou, 15780, Greece
pelekis.chr@gmail.com