This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: Center for Systems and Control, College of Engineering, Peking University, Beijing 100871, P. R. China
School of Mathematical Sciences, University of Electronic Science and Technology of China, Chengdu 611731, P. R. China
longwang@pku.edu.cn

Decision theory and game theory Dynamics of evolution Stochastic modeling

Role of the effective payoff function in evolutionary game dynamics

Feng Huang 11    Xiaojie Chen 22    Long Wang 11**1122**
Abstract

In most studies regarding evolutionary game dynamics, the effective payoff, a quantity that translates the payoff derived from game interactions into reproductive success, is usually assumed to be a specific function of the payoff. Meanwhile, the effect of different function forms of effective payoff on evolutionary dynamics is always left in the basket. With introducing a generalized mapping that the effective payoff of individuals is a non-negative function of two variables on selection intensity and payoff, we study how different effective payoff functions affect evolutionary dynamics in a symmetrical mutation-selection process. For standard two-strategy two-player games, we find that under weak selection the condition for one strategy to dominate the other depends not only on the classical σ\sigma-rule, but also on an extra constant that is determined by the form of the effective payoff function. By changing the sign of the constant, we can alter the direction of strategy selection. Taking the Moran process and pairwise comparison process as specific models in well-mixed populations, we find that different fitness or imitation mappings are equivalent under weak selection. Moreover, the sign of the extra constant determines the direction of one-third law and risk-dominance for sufficiently large populations. This work thus helps to elucidate how the effective payoff function as another fundamental ingredient of evolution affect evolutionary dynamics.

pacs:
02.50.Le
pacs:
87.23.Kg
pacs:
87.10.Mn

1 Introduction

In a Darwinian evolutionary process, it mainly incorporates three fundamental ingredients: inheritance, mutation, and selection. Due to the influence of perturbation or random drift, a genetic process frequently accompanies the occurrence of mutations, which creates genotypic or phenotypic variation, thereby finally leading to the differences in individual fitness that selection acts upon [1]. By integrating game theory with Darwinian evolution [2, 3], evolutionary game theory has become a powerful mathematical framework to model biological [4, 5] and social [6, 7, 8, 9] evolution in a population consisting of different types of interacting individuals under frequency-dependent selection.

Traditionally, a widely used system that focuses on the effects of frequency-dependent selection is the replicator equation [10, 11], where the population is infinitely large well-mixed and the stochastic effect is exclusively overlooked usually. However, if we relax this setting to a more realistic situation where the population is finite well-mixed and subject to fluctuations, this deterministic approach is augmented and disturbed by random drift [12, 13, 14]. In such a finite population with fluctuations, it needs to resort to the tool of stochastic evolutionary game dynamics for investigating the evolution of different traits [1, 15]. In addition to the classical Wright-Fisher process [12, 16], frequency-dependent Moran process [13, 15], and imitation-based pairwise comparison process [17, 18, 19] are two most common microscopic models of strategy spreading in finite populations. In contrast to the well-mixed population setting above, there are also lots of interest in studying evolutionary game dynamics in structured populations [17, 20, 21, 22]. Typically, the spatial geometry of population structure is modeled by regular lattices [23, 24, 25, 26, 27, 28] or more general complex networks [29, 30, 31, 32], where individual interactions merely occur among nearest neighbors.

For a game system, in general, the ingredients influencing the final evolutionary outcomes are nothing but the model, update rule, mutation rate, and population structure, etc. Models and update rules determine the way of strategy spreading. Mutation rates measure the intensity of randomness, while the underlying population structure describes the geometry of individual interactions. Depending on the game interactions, each individual obtains a payoff, and finally it needs to translate into reproductive success, termed effective payoff [33]. For example, the effective payoff is known as fitness in Moran process [13, 15] and imitation probability in pairwise comparison process [18, 19, 34]. Based on the usual assumption that the effective payoff is the form of 1+(Selection intensity)(Payoff1+(\text{Selection intensity})\cdot(\text{Payoff}), for standard two-strategy two-player games, Tarnita et al. demonstrate that if the selection intensity is weak the condition for one strategy to dominate the other is determined by a ‘σ\sigma-rule’ [33], which holds on any population structure [35]. The parameter σ\sigma, termed structure coefficient, is a quantity that only depends on the population structure, update rule, and mutation rate, but not on the payoff values. Later, this work attracts wide interest [36, 37, 38]. For two-player games with multiple strategies, it involves two structure coefficients [36]. To calculate them, investigating games with three strategies is enough. While for multi-player games with two strategies where dd individuals are selected to play a game, the σ\sigma-rule will depend on d1d-1 structure coefficients [37]. In particular, for a more general setting of multi-player games with many strategies, this rule turns out to be quite complicated and the number of structure coefficients required for a symmetric game with dd-player and nn-strategy grows in dd like dn1d^{n-1} [38]. Clearly, because the form of the effective payoff is a specific function in these works, the σ\sigma-rule does not reflect the influence of the effective payoff function on evolutionary outcomes.

In evolutionary biology, however, how to measure the genotype-fitness map (i.e., the fitness landscape) is always a challenging issue, and now it has been accepted that the shape of the genotype-fitness map has fundamental effects on the course of evolution [39]. In addition, based on a Markov chain model, it has been demonstrated that the heterogeneity of individual background fitness can act as a suppressor of selection [40]. In a way, therefore, it means that the form of the effective payoff function (which translates the payoff derived from game interactions into the ability of reproductive success) has a significant effect on the evolution of game dynamics.

To this end, in this letter we study the effect of different function forms of effective payoff on evolutionary dynamics and accordingly extend the results given by Tarnita et al. [33]. We find that if the first-order derivative of the effective payoff function can be written by a linear combination of payoff, then the condition for one strategy to dominate the other depends not only on the σ\sigma-rule, but also on an extra constant which is determined by the effective payoff function. This constant determines the direction of σ\sigma-rule (strategy selection). Additionally, taking the Moran process and pairwise comparison process as specific models in well-mixed populations, we demonstrate that different fitness or imitation mappings are equivalent under weak selection and the extra constant curbs the direction of one-third law and risk-dominance in the limit of large populations.

2 Model and results

In a structured population with NN individuals, we consider stochastic evolutionary dynamics induced by a mutation-selection process. Each player can choose an arbitrary strategy from AA and BB. Then, depending on the payoff matrix

ABA( ab) Bcd,\bordermatrix{~&A&B\cr A&a&b\cr B&c&d\cr}, (1)

players obtain an accumulative payoff by interacting with other individuals based on the underlying population structure. For example, when an AA player interacts with another AA player, it will obtain a payoff aa, but bb when interacting with a BB player. Likewise, a BB player can obtain a payoff cc when interacting with an AA player, and payoff dd when interacting with another BB player. Therefore, the total payoff of each player is a linear function of aa, bb, cc, and dd without including constant terms (evidently, if the payoff is calculated in an average way, the linear relation also holds). For an AA player, for instance, the total payoff is a(number of A-neighbors)+b(number of B-neighbors)a\cdot(\text{number of {A}-neighbors})+b\cdot(\text{number of {B}-neighbors}). To study the effect of effective payoff functions on evolutionary dynamics, instead of a specific form, we assume that the effective payoff of a player is given by φ(β,Payoff)\varphi(\beta,\text{Payoff}). Parameter β\beta measures the intensity of selection, and β0\beta\rightarrow 0 corresponds to the case of weak selection [13, 19, 41].

The reproductive process of each player is dependent on the update rule and its effective payoff, and subject to mutations. With probability μ\mu, a mutation occurs and the offspring adopts a strategy (AA or BB) at random. Otherwise, with probability 1μ1-\mu, the offspring inherits its parent’s strategy. For μ=1\mu=1, there are only mutations, no selection, and strategy choice is completely random. If 0<μ<10<\mu<1, however, there exists a mutation-selection equilibrium [42, 43].

For the game of two strategies, the frequency of AA players in the population defines a finite state space, SS, and the evolutionary dynamics can be captured by a Markov process on this state space. We denote the transition probability from state SiS_{i} to state SjS_{j} by PijP_{ij}. Since the transition probability depends on the update rule and on the effective payoff of players, it can be given by Pij[φ(β,Payoff)]P_{ij}[\varphi(\beta,\text{Payoff})]. Furthermore, we assume that φ(β,Payoff)\varphi(\beta,\text{Payoff}) is differentiable at β=0\beta=0. In the limit of weak selection, then we can give φ(β,Payoff)\varphi(\beta,\text{Payoff}) in the form of first-order Taylor expansions, φ(β,Payoff)=φ0+φ(1)(0)β+o(β)\varphi(\beta,\text{Payoff})=\varphi_{0}+\varphi^{(1)}(0)\cdot\beta+o(\beta). Here φ0:=φ(0,Payoff)\varphi_{0}:=\varphi(0,\text{Payoff}) represents the baseline effective payoff of each player, and φ(1)(0):=[φ(β,Payoff)/β]β=0\varphi^{(1)}(0):=[\partial\varphi(\beta,\text{Payoff})/\partial\beta]_{\beta=0} represents the first-order coefficient of selection intensity. Particularly, if φ(1)(0)\varphi^{(1)}(0) can be written by a linear combination of payoff, that is φ(1)(0)=[φ(β,Payoff)/β]β=0=k0Payoff+c0\varphi^{(1)}(0)=[\partial\varphi(\beta,\text{Payoff})/\partial\beta]_{\beta=0}=k_{0}\cdot\text{Payoff}+c_{0}, then the transition probability is given by Pij[φ0+(k0Payoff+c0)β+o(β)]P_{ij}[\varphi_{0}+(k_{0}\cdot\text{Payoff}+c_{0})\beta+o(\beta)]. Clearly, the constants, k0k_{0} and c0c_{0}, are dependent on the choice of the function φ(β,Payoff)\varphi(\beta,\text{Payoff}) and may rely on the entries of the payoff matrix. In addition, actually a large body of functions meet the condition φ(1)(0)=k0Payoff+c0\varphi^{(1)}(0)=k_{0}\cdot\text{Payoff}+c_{0}, such as 1β+βPayoff1-\beta+\beta\cdot\text{Payoff} [13], exp(βPayoff)exp(\beta\cdot\text{Payoff}) [37], and φ(βPayoff)\varphi(\beta\cdot\text{Payoff}) [19]. Therefore, this condition is not a harsh requirement. Note that the payoff of players is linear in aa, bb, cc, and dd without constant terms, it follows that the transition probability is the function Pij[(k0a+c0)β,(k0b+c0)β,(k0c+c0)β,(k0d+c0)β]P_{ij}[(k_{0}a+c_{0})\beta,(k_{0}b+c_{0})\beta,(k_{0}c+c_{0})\beta,(k_{0}d+c_{0})\beta]. Then, based on the notation (a,b,c,d):=(k0a+c0,k0b+c0,k0c+c0,k0d+c0)(a^{\prime},b^{\prime},c^{\prime},d^{\prime}):=(k_{0}a+c_{0},k_{0}b+c_{0},k_{0}c+c_{0},k_{0}d+c_{0}) and following the proof given in Ref. [33], we know that the condition that strategy AA is favored over strategy BB (i.e., strategy AA is more abundant than BB in the mutation-selection equilibrium) is σa+b>c+σd\sigma a^{\prime}+b^{\prime}>c^{\prime}+\sigma d^{\prime}, where σ\sigma is a parameter that depends on the population structure, update rule, and mutation rate. Accordingly, we have the following theorem:
Theorem 11. Consider a population structure and an update rule that satisfy the following three conditions: (i) the transition probabilities are differentiable at β=0\beta=0; (ii) the update rule is symmetric for the two strategies AA and BB; and (iii) in the game given by the matrix entries, a=c=d=0a=c=d=0 and b=1b=1, strategy AA is not disfavored. Then, in the limit of weak selection, when the effective payoff function φ(β,Payoff)\varphi(\beta,\text{Payoff}) satisfies [φ(β,Payoff)/β]β=0=k0Payoff+c0[\partial\varphi(\beta,\text{Payoff})/\partial\beta]_{\beta=0}=k_{0}\cdot\text{Payoff}+c_{0}, strategy AA is favored over strategy BB if k0(σa+bcσd)>0,k_{0}(\sigma a+b-c-\sigma d)>0, where k0k_{0} is a constant that relies on the function of the effective payoff, and σ\sigma is the structure coefficient which depends on the model and the dynamics (population structure, update rule, and mutation rate), but not on the entries of the payoff matrix.

This theorem implies that for determining the condition under which one strategy dominates the other, the classical σ\sigma-rule [33] is not enough. It also depends on an additional constant k0k_{0} determined by the effective payoff function. Actually, the constant k0k_{0} controls the direction of σ\sigma-rule (strategy selection). If the effective payoff function is given such that the constant k0k_{0} is positive, the theorem recovers the classical σ\sigma-rule (selection favors AA to dominate BB). Otherwise, if the effective payoff function is given such that k0k_{0} is negative, the classical σ\sigma-rule will reverse the direction (selection favors BB to dominate AA).

3 Moran and pairwise comparison process

To check the validity of our theorem and to study how the effective payoff function influences the evolutionary outcomes in a specific dynamic process, here we consider the frequency-dependent Moran process and pairwise comparison process. These two processes represent two classes of typical evolutionary dynamics. The former describes how successful strategies spread in the population through genetic reproduction, whereas the latter describes such a process through cultural imitation.

In the Moran process, the effective payoff is known as the individual fitness, which measures the ability to survive and produce offspring. With a probability proportional to the fitness, an individual is selected randomly for reproduction. And then one identical offspring replaces another randomly chosen individual. Usually, the fitness is assumed to be a convex combination of a background fitness (which is set to one) and the payoff from the game [13, 15], or an exponential function of payoff [37]. Under these specific forms, the constant k0k_{0} related to the fitness function actually turns to 11 and it reduces to the previous results [13, 18]. Instead of using a specific fitness form, here we adopt a generalized mapping that the fitness of a player is any a non-negative function of two variables on selection intensity and payoff, f(β,Payoff)f(\beta,\text{Payoff}). Since mutation occurs during the process of reproduction, it follows that the transition probabilities are given by

Pi,i+1=if(β,πA)(1μ)+(Ni)f(β,πB)μif(β,πA)+(Ni)f(β,πB)NiN,Pi,i1=(Ni)f(β,πB)(1μ)+if(β,πA)μif(β,πA)+(Ni)f(β,πB)iN,\begin{split}P_{i,i+1}&=\frac{if(\beta,\pi_{A})(1-\mu)+(N-i)f(\beta,\pi_{B})\mu}{if(\beta,\pi_{A})+(N-i)f(\beta,\pi_{B})}\frac{N-i}{N},\\ P_{i,i-1}&=\frac{(N-i)f(\beta,\pi_{B})(1-\mu)+if(\beta,\pi_{A})\mu}{if(\beta,\pi_{A})+(N-i)f(\beta,\pi_{B})}\frac{i}{N},\end{split} (2)

where πA(i):=[a(i1)+b(Ni)]/(N1)\pi_{A}(i):=[a(i-1)+b(N-i)]/(N-1) and πB(i):=[ci+d(Ni1)]/(N1)\pi_{B}(i):=[ci+d(N-i-1)]/(N-1) are the average payoffs of an AA player and a BB player, respectively.

While in the pairwise comparison process, the effective payoff is known as the imitation probability. Two individuals are sampled randomly and then a focal player imitates the strategy of the role model with a probability depending on the payoff comparison [43, 18]. As usual, the imitation probability is modeled by the Fermi function with considering the effect of noise [24, 17, 25]. Thus, this process is also called Fermi process. Under the situation that the effective payoff mapping is non-specified, however, the imitation probability function for the pairwise comparison process should be given by g(β,Δπ)g(\beta,\Delta\pi). Here, Δπ\Delta\pi denotes the difference of average payoffs between strategy AA and BB. In the presence of mutations, this imitation process occurs accurately with probability 1μ1-\mu. Otherwise, with probability μ\mu, the focal player adopts a random strategy, AA or BB. Then, it leads to the transition probabilities,

Pi,i+1=(1μ)iNNiNg(β,Δπ(i))+NiNμ2,Pi,i1=(1μ)iNNiNg(β,Δπ(i))+iNμ2,\begin{split}P_{i,i+1}&=(1-\mu)\frac{i}{N}\frac{N-i}{N}g(\beta,\Delta\pi(i))+\frac{N-i}{N}\frac{\mu}{2},\\ P_{i,i-1}&=(1-\mu)\frac{i}{N}\frac{N-i}{N}g(\beta,-\Delta\pi(i))+\frac{i}{N}\frac{\mu}{2},\end{split} (3)

where Δπ(i):=πA(i)πB(i)=ui+v\Delta\pi(i):=\pi_{A}(i)-\pi_{B}(i)=ui+v. Herein, parameters uu and vv are defined by u:=(abc+d)/(N1)u:=(a-b-c+d)/(N-1) and v:=(NbNda+d)/(N1)v:=(Nb-Nd-a+d)/(N-1), respectively.

Moreover, for both processes, the probability to stay in the current state is 1Pi,i+1Pi,i11-P_{i,i+1}-P_{i,i-1}, and the probability to transform to other states is vanishing. In what follows, we first calculate the fixation probabilities and fixation times under weak selection when mutations are absent. Then, we derive the criterion that strategy AA is favored over strategy BB in this case, and finally extrapolate this criterion to small mutation rates.

3.1 Fixation probabilities and fixation times

If there are no mutations in these two game systems, then one quantity of most interest is the fixation probability, ϕi\phi_{i}, which describes the probability that ii individuals of type AA reach fixation at all AA. Another significant quantity is the average time for a single AA player reaching fixation [44, 45, 46]. The former measures the preference of natural selection whereas the latter characterizes the evolutionary velocity of the system.

First, we follow the conditions given by Theorem 11, that is, the first-order derivative of fitness function for the Moran process can be written by fβ(0,π)=k0(m)π+c0(m)f_{\beta}(0,\pi)=k_{0}^{(m)}\cdot\pi+c_{0}^{(m)}, and the one of imitation probability function for the pairwise comparison process can be written by gβ(0,Δπ)=k0(p)Δπ+c0(p)g_{\beta}(0,\Delta\pi)=k_{0}^{(p)}\cdot\Delta\pi+c_{0}^{(p)}. Here, k0(m)k_{0}^{(m)} and c0(m)c_{0}^{(m)} (k0(p)k_{0}^{(p)} and c0(p)c_{0}^{(p)}) are two constants that depend on the choice of fitness (imitation probability) functions and may be related to the entries of the payoff matrix. With the notations m0:=k0(m)/f0m_{0}:=k_{0}^{(m)}/f_{0} and p0:=2k0(p)/g0=k0(p)/g02p_{0}:=2k_{0}^{(p)}/g_{0}=k_{0}^{(p)}/g_{0}^{2}, where f0=f(0,π)f_{0}=f(0,\pi) is the baseline fitness of each player and g0=g(0,Δπ)=1/2g_{0}=g(0,\Delta\pi)=1/2 is the probability of random imitation, we obtain the approximation of fixation probabilities under weak selection (details for the Supplementary Material) as

ϕiiN+βsi(Ni)[(N+i)u+3v]6N,\phi_{i}\approx\frac{i}{N}+\beta\cdot s\cdot\frac{i(N-i)[(N+i)u+3v]}{6N}, (4)

where s=m0s=m_{0} for the frequency-dependent Moran process, and s=p0s=p_{0} for the pairwise comparison process.

While for the average times of a single AA player reaching fixation, there are two kinds of fixation times that attract much research attention [44, 18]. The first one is the unconditional average time of fixation t1t_{1}, which is the expected value for the time until the population reaches one of the two absorbing states, all AA and all BB, when starting from a single AA. Another is the conditional average time of fixation t1At_{1}^{A}, which specifies the expected time that a player of type AA takes to reach the absorbing state, all AA. In the limit of weak selection, we find that the unconditional and conditional fixation times for the Moran process (details for the Supplementary Material) can be approximated to

t1NHN1+m0vN2(N+12HN)β,t1AN(N1)m0uN2(N23N+2)36β,\begin{split}t_{1}&\approx NH_{N-1}+m_{0}v\frac{N}{2}(N+1-2H_{N})\beta,\\ t_{1}^{A}&\approx N(N-1)-m_{0}u\frac{N^{2}(N^{2}-3N+2)}{36}\beta,\end{split} (5)

whereas for the pairwise comparison process, they are given by

t12NHN1+p0vN(N1HN1)β,t1A2N(N1)p0uN(N1)N2+N618β,\begin{split}t_{1}\approx 2NH_{N-1}+p_{0}vN(N-1-H_{N-1})\beta,\\ t_{1}^{A}\approx 2N(N-1)-p_{0}uN(N-1)\frac{N^{2}+N-6}{18}\beta,\end{split} (6)

where HN=l=1N1lH_{N}=\sum_{l=1}^{N}\frac{1}{l} is the harmonic number.

Interestingly, for both Moran process and pairwise comparison process, if the first-order derivative of the effective payoff function (i.e., the fitness and imitation probability function) can be written by a linear combination of the payoff, the difference in the influence of effective payoff functions on evolutionary outcomes just embodies in the coefficients before selection intensity, m0m_{0} and p0p_{0}. By proper rescaling, actually, these constant coefficients can be absorbed into the selection intensity, or make all payoff matrix entries (aa, bb, cc, and dd) change a scale in view of the exact formulae of uu and vv. In particular, if we adopt a linear or an exponential form of payoff as the fitness function, or the Fermi function as the imitation probability, both m0m_{0} and p0p_{0} are 11, which recovers the previous results [18] as specific cases. Moreover, under the conditions [(N+1)u+3v]>0[(N+1)u+3v]>0 and u>0u>0, if a fitness (imitation probability) function is chosen such that m0>0m_{0}>0 (p0>0p_{0}>0), then taking the constant 11 as the benchmark, m0>1m_{0}>1 (p0>1p_{0}>1) leads this function to acting as an amplifier of selection (facilitating the fixation of advantage individuals and decreasing the fixation time). Nevertheless, when m0<1m_{0}<1 (p0<1p_{0}<1), this function acts as a suppressor of selection (suppressing the fixation of advantage individuals and increasing the fixation time). This result holds not only for weak selection, but also for intermediate selection intensity (see Fig. 1). As the counterpart, if the function is chosen such that m0<0m_{0}<0 (p0<0p_{0}<0), with the scaling theory of dilemma strength [47, 48], another example of prisoner’s dilemma is given in the Supplementary Material.

3.2 Equivalence

Based on the above calculations of fixation probabilities and fixation times under weak selection, additionally, we find that two arbitrary fitness (imitation probability) functions in a Moran process (pairwise comparison process) are equivalent. Specifically, for the frequency-dependent Moran process with a generalized fitness function f(β,π)f(\beta,\pi), if fβ(0,π)=k0(m)π+c0(m)f_{\beta}(0,\pi)=k_{0}^{(m)}\cdot\pi+c_{0}^{(m)} is satisfied, we know that the influence of any two different fitness functions on evolutionary outcomes just embodies in the constant factor m0m_{0} before the selection intensity. Thus, in this sense, any two fitness mappings meeting the conditions defined above are equivalent under weak selection. The equivalence means that the difference in fixation probabilities and fixation times is merely captured by the constant factor m0m_{0}, with which the payoff matrix or the intensity of selection changes a scale.

Particularly, if the fitness function adopts one of the function families, F1(β,π)=j=0majβj+i=1nbi(βπ)iF_{1}(\beta,\pi)=\sum_{j=0}^{m}a_{j}\beta^{j}+\sum_{i=1}^{n}b_{i}(\beta\pi)^{i} and F2(β,π)=j=0majβj+i=1nbiβiπ(m,n=1,2,3,)F_{2}(\beta,\pi)=\sum_{j=0}^{m}a_{j}\beta^{j}+\sum_{i=1}^{n}b_{i}\beta^{i}\pi\ (m,n=1,2,3,...), where aja_{j} and bib_{i} are constant coefficients that guarantee F1>0F_{1}>0 and F2>0F_{2}>0 because the individual fitness is positive [37, 19], then we have the same factor m0=b1/a0m_{0}=b_{1}/a_{0}. Interestingly, if m0=b1/a0=1m_{0}=b_{1}/a_{0}=1, these two function families are equivalent to the prevalent fitness mappings 1β+βπ1-\beta+\beta\pi and exp(βπ)exp(\beta\pi) under weak selection (see Fig. 2). Actually, the Taylor series of exp(βπ)exp(\beta\pi) at β=0\beta=0 is just the function family F1F_{1} when specific coefficients a0=1a_{0}=1, aj=0(j>0)a_{j}=0\ (j>0), and bi=1/(i!)b_{i}=1/(i!) are applied.

Similarly, for the pairwise comparison process with a generalized imitation probability function g(β,Δπ)g(\beta,\Delta\pi), if gβ(0,Δπ)=k0(p)Δπ+c0(p)g_{\beta}(0,\Delta\pi)=k_{0}^{(p)}\cdot\Delta\pi+c_{0}^{(p)} is satisfied, then the influence of any two different imitation probability functions on evolutionary outcomes also embodies in a constant factor before the selection intensity, p0p_{0}. Thus, in this sense, for two arbitrary imitation probability functions, they are also equivalent under weak selection. The equivalence follows the same meaning as that in the Moran process, that is, the constant factor p0p_{0} uniquely measures the difference of fixation probabilities and fixation times under weak selection.

Surprisingly, with a completely similar formulation to the fitness function families F1F_{1} and F2F_{2}, G1(β,Δπ)=j=0mαjβj+i=1nηi(βΔπ)iG_{1}(\beta,\Delta\pi)=\sum_{j=0}^{m}\alpha_{j}\beta^{j}+\sum_{i=1}^{n}\eta_{i}(\beta\Delta\pi)^{i} and G2(β,Δπ)=j=0mαjβj+i=1nηiβiΔπ(m,n=1,2,3,)G_{2}(\beta,\Delta\pi)=\sum_{j=0}^{m}\alpha_{j}\beta^{j}+\sum_{i=1}^{n}\eta_{i}\beta^{i}\Delta\pi\ (m,n=1,2,3,...), where αj\alpha_{j} and ηi\eta_{i} are constant coefficients which guarantee that G1G_{1} and G2G_{2} are probability functions and α0=1/2\alpha_{0}=1/2, are two classes of equivalent imitation probability functions. Particularly, if η1=1/4\eta_{1}=1/4, these two function families are equivalent to the popular Fermi function 1/(1+eβΔπ)1/(1+e^{-\beta\Delta\pi}) under weak selection (see Fig. 2). Actually, by choosing specific coefficients, G1G_{1} can also become the Taylor series of Fermi function at β=0\beta=0.

Refer to caption
Figure 1: Effect of fitness and imitation probability functions on evolutionary dynamics acts as a selection amplifier or suppressor. Using the payoff matrix entries, a=1.2a=1.2, b=0.8b=0.8, c=1.0c=1.0, and d=0.7d=0.7 (i.e., AA dominates BB), we show the fixation probabilities and conditional fixation times of a single AA when population size is N=100N=100 in Moran process (panel (a) and (b)) and pairwise comparison process (panel (c) and (d)), respectively. Lines are analytical results (Eqs. (S99) and (S1414) in the Supplementary Material), while symbols are simulations. If a fitness function is applied such that m0>1m_{0}>1 ((a) and (b)) or an imitation probability function such that p0>1p_{0}>1 ((c) and (d)), it promotes the fixation of advantage strategy and decreases the fixation time (red lines) compared with the benchmark m0=1m_{0}=1 or p0=1p_{0}=1 (blue lines). Otherwise, if m0<1m_{0}<1 or p0<1p_{0}<1, it suppresses the fixation of advantage strategy and increases the fixation time (purple lines) compared with the benchmark (blue lines).
Refer to caption
Figure 2: Equivalence of Moran process and pairwise comparison process under weak selection. In the first row, we show the fixation probability ϕ1\phi_{1} (a), unconditional fixation time t1t_{1} (b), and conditional fixation time t1At_{1}^{A} (c) of a single mutant AA in Moran process, respectively. The exact analytical results are depicted by solid lines with symbols, and accordingly we show the weak selection approximations by short dash dots, based on a class of equivalent fitness functions. The same manipulation is repeated for pairwise comparison process with a class of equivalent imitation probability functions, as shown in the second row. Analytical results are numerical calculations on the basis of exact Eqs. (S99), (S1212), and (S1414) in the Supplementary Material, but weak selection approximations are based on Eqs. (4)–(6). Parameters are N=100N=100, a=4a=4, b=1b=1, c=1c=1, and d=5d=5 (coordination games) in all panels. (These results are also valid for dominance and coexistence games.)

3.3 One-third law and risk-dominance

In stochastic evolutionary game dynamics, the notions of invasion and fixation are two fundamental concepts to describe the spreading of strategies in finite populations [13, 15]. Using the neutral game as the benchmark, strategy AA is shortly said to fixate in a resident population (selection favors AA replacing BB) if the fixation probability for a single AA is larger than that in the neutral game [13, 15]. Thus, for the frequency-dependent Moran process with a fitness function f(β,π)f(\beta,\pi) which fulfills the condition given above, selection favors AA replacing BB under weak selection if m0[(N+1)u+3v]>0m_{0}[(N+1)u+3v]>0 (see Eq. (4)). Similarly, for the pairwise comparison process with an imitation probability function g(β,Δπ)g(\beta,\Delta\pi), selection favors AA replacing BB under weak selection if p0[(N+1)u+3v]>0p_{0}[(N+1)u+3v]>0 (see Eq. (4)). In view of the notations m0:=k0(m)/f0m_{0}:=k_{0}^{(m)}/f_{0} and p0:=2k0(p)/g0=k0(p)/g02p_{0}:=2k_{0}^{(p)}/g_{0}=k_{0}^{(p)}/g_{0}^{2}, it follows that the criterion that selection favors AA replacing BB under weak selection is k0[(N+1)u+3v]>0k_{0}[(N+1)u+3v]>0, where k0=k0(m)k_{0}=k_{0}^{(m)} for the Moran process and k0=k0(p)k_{0}=k_{0}^{(p)} for the pairwise comparison process. Specifically, we have (see Fig. 3): (i) When k0>0k_{0}>0, the condition under which selection favors AA replacing BB (ϕ1>1/N\phi_{1}>1/N) is a(N2)+b(2N1)>c(N+1)+d(2N4)a(N-2)+b(2N-1)>c(N+1)+d(2N-4). Particularly, for sufficiently large population size N+N\rightarrow+\infty, it corresponds to one-third law [13] in the case of coordination games (1/3>(db)/(abc+d)1/3>(d-b)/(a-b-c+d)); (ii) When k0<0k_{0}<0, this condition changes to a(N2)+b(2N1)<c(N+1)+d(2N4)a(N-2)+b(2N-1)<c(N+1)+d(2N-4). Particularly, for N+N\rightarrow+\infty, the classical one-third law is reversed in the case of coordination games (1/3<(db)/(abc+d)1/3<(d-b)/(a-b-c+d)); (iii) When k0=0k_{0}=0, the condition that selection favors AA replacing BB will depend on the higher order coefficients of β\beta in ϕ1\phi_{1}. Actually, the calculations of higher order coefficients under weak selection are more tedious than the linear approximation [19].

Except for the underlying principle that determines the condition of favoring strategy AA to replace BB, it is also of interest to ask whether strategy AA is selected over strategy BB, termed ‘strategy selection’ [33]. First, let ρA\rho_{A} (ρB\rho_{B}) denotes the fixation probability that a single individual using strategy AA (BB) invades and takes over a resident population of BB (AA) players. Accordingly, we have ρA=ϕ1\rho_{A}=\phi_{1}. Moreover, note that the probability ρB\rho_{B} is equal to that N1N-1 individuals of type AA fail to take over a population in which there is just a single BB individual. Then, it leads to ρB=1ϕN1\rho_{B}=1-\phi_{N-1}. With introducing the notation γj=Pj,j1/Pj,j+1\gamma_{j}=P_{j,j-1}/P_{j,j+1}, which is the ratio of transition probabilities when mutations are absent (μ=0\mu=0), we have the ratio of fixation probabilities for strategy AA and BB as ρB/ρA=j=1N1γj\rho_{B}/\rho_{A}=\prod_{j=1}^{N-1}\gamma_{j}. Under weak selection, the ratio of these two fixation probabilities can be approximated to

ρB/ρA1sβj=1N1(uj+v)=1sβ2[a(N2)+bNcNd(N2)],\begin{split}\rho_{B}/\rho_{A}&\approx 1-s\beta\sum_{j=1}^{N-1}(uj+v)\\ &=1-\frac{s\beta}{2}[a(N-2)+bN-cN-d(N-2)],\end{split} (7)

where s=m0s=m_{0} for the Moran process, and s=p0s=p_{0} for the pairwise comparison process. In view of the definitions of m0m_{0} and p0p_{0}, therefore, the condition under which strategy AA is selected over strategy BB is given by k0(σa+bcσd)>0k_{0}(\sigma a+b-c-\sigma d)>0, where k0=k0(m)k_{0}=k_{0}^{(m)} for the Moran process and k0=k0(p)k_{0}=k_{0}^{(p)} for the pairwise comparison process, and σ=(N2)/N\sigma=(N-2)/N is the structure coefficient of well-mixed populations [33]. Under weak selection, specifically we have (see Fig. 3): (i) When k0>0k_{0}>0, the condition that strategy AA is selected over strategy BB is σa+b>c+σd\sigma a+b>c+\sigma d. Particularly, for sufficiently large population size N+N\rightarrow+\infty, it corresponds to that AA is risk-dominant in the case of coordination games (1/2>(db)/(abc+d)1/2>(d-b)/(a-b-c+d)); (ii) When k0<0k_{0}<0, this condition changes to σa+b<c+σd\sigma a+b<c+\sigma d. Particularly, for N+N\rightarrow+\infty, it corresponds to that BB is risk-dominant in the case of coordination games (1/2<(db)/(abc+d)1/2<(d-b)/(a-b-c+d)); (iii) When k0=0k_{0}=0, the condition that strategy AA is selected over strategy BB under weak selection will depend on higher order coefficients of β\beta in ρB/ρA\rho_{B}/\rho_{A}.

In particular, if we additionally consider the situation where small non-uniform mutations occur between the two strategies, that strategy AA is more abundant than BB in the long run is determined by μABρB/μBAρA<1\mu_{AB}\rho_{B}/\mu_{BA}\rho_{A}<1 [42, 44, 49], where μAB\mu_{AB} and μBA\mu_{BA} denote the mutation rates from AA to BB and from BB to AA, respectively. For μAB=μBA\mu_{AB}=\mu_{BA}, clearly the conclusions obtained above are still valid, which extrapolates our results to the situation of small mutations.

Refer to caption
Figure 3: Direction of the one-third law and the risk dominance is determined by the choice of fitness and imitation probability functions. When k0(m)>0k_{0}^{(m)}>0 for Moran process or k0(p)>0k_{0}^{(p)}>0 for pairwise comparison process, it corresponds to the classical one-third law and risk-dominance in coordination games (a). While if k0(m)<0k_{0}^{(m)}<0 for Moran process or k0(p)<0k_{0}^{(p)}<0 for pairwise comparison process, the direction of one-third law and risk-dominance is reversed (b). Otherwise, the conditions determining ρA>1/N\rho_{A}>1/N or ρA>ρB\rho_{A}>\rho_{B} under weak selection will depend on higher order coefficients of fixation probabilities.

4 Discussion

Most previous studies exclusively assume that the effective payoff function is a specific form. One direct result of this setting is that the final evolutionary outcomes do not reflect the effect of the effective payoff function on evolutionary dynamics. Therefore, it still remains unclear how the effective payoff function influences the evolutionary dynamics in a game system. With introducing a generalized mapping, we investigate such effect in this letter.

For standard 2×22\times 2 games where a specific form of the effective payoff is adopted, it has been demonstrated that under weak selection the condition for one strategy to dominate the other is determined by a σ\sigma-rule [33]. This rule almost captures all aspects of evolutionary dynamics, but ignores the effect of the effective payoff function. Particularly, if the effective payoff function is an any non-negative mapping of the product of the payoff and the selection intensity [19, 37], the rule still holds. But it does not change the basic fact that the role played by the effective payoff function in evolutionary game dynamics is exclusively overlooked. With introducing a more generalized mapping that the effective payoff of individuals is a function of payoff and selection intensity, however, we find that the condition determining a strategy to be selected relies not only on the σ\sigma-rule, but also on an extra constant which characterizes and depends on the effective payoff function. As an extension, a multi-strategy version is also given in the Supplementary Material.

Based on specific effective payoff forms, it has been found that the linear and exponential functions lead to identical evolutionary outcomes under weak selection [43, 37, 19]. Here we generalize this equivalence understanding to any two fitness functions in a Moran process, and imitation probability functions in a pairwise comparison process. In addition, except for the standard 2×22\times 2 games and weak selection, there are lots of research interest in the games of multiple players or strategies [37, 50, 36] and strong selection [51], which are worth the effort in the future.

Acknowledgements.
We thank Bin Wu for helpful discussions and comments. This work was supported by the National Natural Science Foundation of China (Grants No. 61751301, No. 61533001, and No. 61503062).

References

  • [1] \NameNowak M. A. \BookEvolutionary Dynamics: Exploring the Equations of Life (Harvard University Press, Cambridge MA) 2006.
  • [2] \NameMaynard Smith J. Price G. R. \REVIEWNature246197315.
  • [3] \NameMaynard Smith J. \BookEvolution and the Theory of Games (Cambridge University Press, Cambridge) 1982.
  • [4] \NameTurner P. E. Chao L. \REVIEWNature3981999441.
  • [5] \NameKerr B., Riley M. A., Feldman M. W. Bohannan B. J. \REVIEWNature4182002171.
  • [6] \NameHamilton W. D. \REVIEWJ. Theor. Biol.719641.
  • [7] \NameNowak M. A. Sigmund K. \REVIEWNature43720051291.
  • [8] \NameHelbing D. Yu W. \REVIEWProc. Natl. Acad. Sci. U.S.A.10620093680.
  • [9] \NameChen X. Szolnoki A. \REVIEWPLoS Comput. Biol.142018e1006347.
  • [10] \NameTaylor P. D. Jonker L. B. \REVIEWMath. Biosci.401978145.
  • [11] \NameHofbauer J. Sigmund K. \BookEvolutionary Games and Population Dynamics (Cambridge University Press, Cambridge) 1998.
  • [12] \NameVan Kampen N. G. \BookStochastic Processes in Physics and Chemistry 2nd Edition (Elsevier, Amsterdam) 1997.
  • [13] \NameNowak M. A., Sasaki A., Taylor C. Fudenberg D. \REVIEWNature4282004646.
  • [14] \NameSui X., Cong R., Li K. Wang L. \REVIEWPhys. Lett. A37920152922.
  • [15] \NameTaylor C., Fudenberg D., Sasaki A. Nowak M. A. \REVIEWBull. Math. Biol.6620041621.
  • [16] \NameImhof L. A. Nowak M. A. \REVIEWJ. Math. Biol.522006667.
  • [17] \NameSzabó G. Fath G. \REVIEWPhys. Rep.446200797.
  • [18] \NameAltrock P. M. Traulsen A. \REVIEWNew J. Phys.112009013012.
  • [19] \NameWu B., Altrock P. M., Wang L. Traulsen A. \REVIEWPhys. Rev. E822010046106.
  • [20] \NameSzolnoki A., Perc M. Danku Z. \REVIEWPhysica A38720082075.
  • [21] \NamePerc M. Szolnoki A. \REVIEWNew J. Phys.142012043013.
  • [22] \NamePerc M., Gómez-Gardeñes J., Szolnoki A., Floría L. M. Moreno Y. \REVIEWJ. R. Soc. Interface10201320120997.
  • [23] \NameNowak M. A. May R. M. \REVIEWNature3591992826.
  • [24] \NameSzabó G. Tőke C. \REVIEWPhys. Rev. E58199869.
  • [25] \NameHauert C. Szabó G. \REVIEWAm. J. Phys.732005405.
  • [26] \NamePerc M. \REVIEWEPL752006841.
  • [27] \NamePerc M. Szolnoki A. \REVIEWBioSystems992010109.
  • [28] \NameChen X., Fu F. Wang L. \REVIEWPhys. Rev. E782008051120.
  • [29] \NameOhtsuki H., Hauert C., Lieberman E. Nowak M. A. \REVIEWNature4412006502.
  • [30] \NameSantos F. C., Santos M. D. Pacheco J. M. \REVIEWNature4542008213.
  • [31] \NameFu F., Wang L., Nowak M. A. Hauert C. \REVIEWPhys. Rev. E792009046707.
  • [32] \NameChen X. Wang L. \REVIEWPhys. Rev. E772008017103.
  • [33] \NameTarnita C. E., Ohtsuki H., Antal T., Fu F. Nowak M. A. \REVIEWJ. Theor. Biol.2592009570.
  • [34] \NameHuang F., Chen X. Wang L. \REVIEWSci. Rep.82018528.
  • [35] \NameAllen B., Lippner G., Chen Y.-T., Fotouhi B., Momeni N., Yau S.-T. Nowak M. A. \REVIEWNature5442017227.
  • [36] \NameTarnita C. E., Wage N. Nowak M. A. \REVIEWProc. Natl. Acad. Sci. U.S.A.10820112334.
  • [37] \NameWu B., Traulsen A. Gokhale C. S. \REVIEWGames42013182.
  • [38] \NameMcAvoy A. Hauert C. \REVIEWJ. Math. Biol.722016203.
  • [39] \NameDe Visser J. A. G. Krug J. \REVIEWNature Rev. Genet.152014480.
  • [40] \NameHauser O. P., Traulsen A. Nowak M. A. \REVIEWJ. Theor. Biol.3432014178.
  • [41] \NameWu B., García J., Hauert C. Traulsen A. \REVIEWPLoS Comput. Biol.92013e1003381.
  • [42] \NameFudenberg D., Nowak M. A., Taylor C. Imhof L. A. \REVIEWTheor. Popul. Biol.702006352.
  • [43] \NameAntal T., Nowak M. A. Traulsen A. \REVIEWJ. Theor. Biol.2572009340.
  • [44] \NameAntal T. Scheuring I. \REVIEWBull. Math. Biol.6820061923.
  • [45] \NameTaylor C., Iwasa Y. Nowak M. A. \REVIEWJ. Theor. Biol.2432006245.
  • [46] \NameSui X., Wu B. Wang L. \REVIEWPhys. Rev. E922015062124.
  • [47] \NameWang Z., Kokubo S., Jusup M. Tanimoto J. \REVIEWPhys. Life Rev.14201556.
  • [48] \NameTanimoto J. Sagara H. \REVIEWBioSystems902007105.
  • [49] \NameWu B., Gokhale C. S., Wang L. Traulsen A. \REVIEWJ. Math. Biol.642012803.
  • [50] \NameSu Q., Wang L. Stanley H. E. \REVIEWNew J. Phys.202018103030.
  • [51] \NameAltrock P. M., Traulsen A. Nowak M. A. \REVIEWPhys. Rev. E952017022407.