Next Article in Journal
Characterizing Actions in a Dynamic Common Pool Resource Game
Next Article in Special Issue
Categorization and Cooperation across Games
Previous Article in Journal
This Is How We Do It: How Social Norms and Social Identity Shape Decision Making under Uncertainty
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evolution of Groupwise Cooperation: Generosity, Paradoxical Behavior, and Non-Linear Payoff Functions

1
Department of Biological Sciences, The University of Tokyo, Hongo 7-3-1, Bunkyo-ku, Tokyo 113-0033, Japan
2
Division of Natural Resource Economics, Graduate School of Agriculture, Kyoto University, Oiwake-cho, Kitashirakawa, Sakyo-ku, Kyoto 606-8502, Japan
3
Key Lab of Animal Ecology and Conservation Biology, Institute of Zoology, Chinese Academy of Science, Datun Road, Chaoyang, Beijing 100101, China
4
School of Economics and Management, Kochi University of Technology, Kochi 780-8515, Japan
5
Meiji Institute for Advanced Study of Mathematical Sciences, Meiji Univeristy, Nakano 4-21-1, Nakano-ku, Tokyo 164-8525, Japan
*
Author to whom correspondence should be addressed.
Games 2018, 9(4), 100; https://doi.org/10.3390/g9040100
Submission received: 29 September 2018 / Revised: 8 November 2018 / Accepted: 13 November 2018 / Published: 10 December 2018
(This article belongs to the Special Issue The Evolution of Cooperation in Game Theory and Social Simulation)

Abstract

:
Evolution of cooperation by reciprocity has been studied using two-player and n-player repeated prisoner’s dilemma games. An interesting feature specific to the n-player case is that players can vary in generosity, or how many defections they tolerate in a given round of a repeated game. Reciprocators are quicker to detect defectors to withdraw further cooperation when less generous, and better at maintaining a long-term cooperation in the presence of rare defectors when more generous. A previous analysis on a stochastic evolutionary model of the n-player repeated prisoner’s dilemma has shown that the fixation probability of a single reciprocator in a population of defectors can be maximized for a moderate level of generosity. However, the analysis is limited in that it considers only tit-for-tat-type reciprocators within the conventional linear payoff assumption. Here we extend the previous study by removing these limitations and show that, if the games are repeated sufficiently many times, considering non-tit-for-tat type strategies does not alter the previous results, while the introduction of non-linear payoffs sometimes does. In particular, under certain conditions, the fixation probability is maximized for a “paradoxical” strategy, which cooperates in the presence of fewer cooperating opponents than in other situations in which it defects.

1. Introduction

Reciprocity is a key factor in the study of evolution of cooperation [1,2]. Evolution of pairwise and groupwise cooperation by reciprocity has been studied by using the two-player and n -player repeated prisoner’s dilemma (PD) games [2,3,4,5,6,7,8,9,10,11,12]. While a population of reciprocators can be stable against invasion by defectors in repeated PDs, so long as games are repeated sufficiently many times, various additional mechanisms have been proposed to explain the initial emergence of reciprocators in a population of defectors, such as invasion by a mass of reciprocators [13], formation of spatial clusters [14,15], stochastic aggregate payoffs [16], and random drift [17]. Of particular relevance to the present study is Nowak et al. (2004). They developed a model of stochastic evolutionary dynamics for the two-player repeated PD and derived the fixation probability with which a single mutant of reciprocator that appears in a population of defectors will eventually replace the whole population. Nowak et al.’s (2004) framework has been extended to include the more general n -player repeated PD [18,19], which has enabled to investigate the evolution of group-wise cooperation [20,21,22] considering the effect of random drift. In this article, we report our further investigation on the stochastic evolutionary model of the n -player repeated PD.
An interesting feature of group-wise cooperation, or more specifically, the n -player repeated PD, is that we can conceive of multiple types of reciprocators that differ in their levels of generosity. Take tit-for-tat (TFT) for example. TFT is a reciprocal strategy in the two-player repeated PD, which cooperates in the first round of a repeated game and, from the second round onward, cooperates if and only if its opponent has cooperated in the previous round. In the n -player repeated PD, n different reciprocal strategies analogous to TFT, or TFTa for a { 0 ,   1 ,   ,   n 1 } , are possible, where a represents the strategy’s level of generosity [23]. Namely, TFTa, cooperates in the first round and after that cooperates if and only if a or more of its opponents have cooperated in the previous round. Thus, TFTa with a smaller value of a is more generous toward the group members’ defection.
Suppose that a single mutant of the strategy that always defects (ALLD) appears in a population of TFTa. Obviously, less generous TFTa (i.e., having larger a ) is quicker to find ALLDs in the group and withdraw further cooperation and, thus, more likely to be stable against their invasion. On the other hand, for TFTa to be advantageous over ALLD, it is always necessary that each TFTa individual tends to have a or more other TFTa individuals in the same group; otherwise, TFTa cooperates only in the first round of each game to be exploited by ALLDs. This means that more generous TFTa (i.e., having smaller a ) may compete better against ALLD, particularly when TFTa is rare in the population; in other words, generosity may serve to facilitate the initial emergence of TFTa.
This intuition has been confirmed by Kurokawa et al. (2010), who compared the fixation probabilities of TFTas with different values of a when appears as a single mutant in a population of ALLD [19]. They found that more generous TFTa can sometimes have a greater fixation probability than less generous ones. They also specified the optimal level of generosity that maximizes the fixation probability, showing that there is a certain threshold level of generosity above which the largest fixation probability can never be attained. Specifically, it was shown that any such TFTa that tolerates defections by a half or more of its opponents can never have the largest fixation probability.
Although Kurokawa et al. (2010) focused on the strategies classified as TFTa, they are not the only ones that can potentially establish mutual cooperation in the n -player repeated PD. In fact, tolerating the presence of some defectors in a group may have a similar enhancing effect on the evolution of reciprocal strategies in general. However, this possibility has not been explored thus far. Here we extend Kurokawa et al.’s (2010) analysis to consider the set of all “reactive strategies” that (i) cooperates with probability d φ { 0 ,   1 } in the first round and (ii) cooperates in the k th round ( k 2 ) with probability d j { 0 ,   1 } , which depends on the number of opponents, j , who have cooperated in the k 1 th round. There are 2 n + 1 such strategies in total, including n 1 TFTas and 2 n strategies that defect in the first move [24,25,26,27,28]. The rest of the strategies in the set are “paradoxical” in the sense that each of them has a certain situation in which it cooperates even though there are fewer cooperators in the group than in other situations in which it withdraws cooperation. It is of interest whether the strategies defecting in the first move or exhibiting paradoxical behavior can have the largest fixation probability among the set of all reactive strategies when invading ALLD.
Another limitation of Kurokawa et al.’s (2010) work is that they only considered the conventional linear payoff functions (i.e., linear public goods games. However, there are at least two reasons why non-linear payoff functions should be also considered. First, in the real world, payoffs are not always linear [29,30,31,32,33,34,35,36,37,38,39] and it has been demonstrated that introduction of non-linearity can affect the outcome of evolutionary models (e.g., [40,41,42,43,44,45,46,47]. Second, if payoffs increase non-linearly with the number of cooperating members in the group, it can be said that the efficiency of an act of cooperation varies depending on the number of cooperators. For example, in case payoffs increase more rapidly when cooperators are fewer, an individual’s act of cooperation is more efficient in the presence of fewer cooperators. In such a case, more generous strategies may be selectively favored because generosity promotes cooperation when it is efficient. To examine these possibilities, we also extend Kurokawa et al. (2010) in terms of payoff functions.
In addition, Kurokawa et al. (2010) derived their results under the assumption that selection is sufficiently weak. However, the intensity of selection can affect the outcomes of evolutionary models (e.g., [17,48,49,50]. Hence, we investigate to what extent our results based on the weak selection assumption may be affected by the selection intensity.
In this paper, we extend Kurokawa et al.’s (2010) analysis by removing these limitations. In particular, taking non-linear payoff functions and paradoxical (i.e., non-tit-for-tat type) strategies into consideration, we ask how non-linear payoffs may facilitate the evolution of excessive generosity and whether any paradoxical strategies can ever attain the highest fixation probability among all reactive strategies considered. Additionally, we examine how our results may depend on the intensity of natural selection. In what follows, we first introduce the n -player repeated PD and our model of stochastic evolutionary dynamics (Model). Then we derive, based on the weak selection approximation, the best reactive strategies, or the strategies maximizing the fixation probability, and provide detailed analyses for linear and non-linear payoff cases. We also examine to what extent our results may be affected if selection is not sufficiently weak (Results). Finally, we summarize the results and discuss some caveats and possibilities of further investigations (Discussion).

2. Model

We investigate stochastic evolutionary dynamics of a population of individuals whose fitness is determined by n -player repeated PD games. We consider a set of “reactive strategies,” as defined below, and compare the fixation probabilities of different reactive strategies when introduced as a mutant into a population of the strategy that always defects, or ALLD. Our goal is to specify the “best reactive strategies,” which maximize the fixation probability.

2.1. The n -Player Repeated Prisoner’s Dilemma

In the n-player repeated PD, a group of n individuals play a game consisting of one or more rounds, in each of which each individual chooses to either cooperate or defect. After one round is finished, another round will be played with probability δ ; otherwise, the group will be dismissed ( 0 < δ < 1 . Thus, the expected number of rounds is 1 / ( 1 δ ) . Suppose that there are k cooperating and n k defecting individuals in a given round. The payoffs to cooperating (C) and defecting (D) individuals in that round are denoted by V ( C | k ) and V ( D | k ) , respectively ( 0 k n . Note that V ( C | 0 ) and V ( D | n ) are defined arbitrarily and used only for the sake of notational convenience. Following Boyd & Richerson (1988), n-player PD requires the following four conditions to be satisfied:
V ( C | k + 1 ) < V ( D | k ) ,
V ( C | k ) < V ( C | k + 1 ) ,
V ( D | k ) < V ( D | k + 1 ) ,
( n k 1 ) V ( D | k + 1 ) + ( k + 1 ) V ( C | k + 1 ) > ( n k ) V ( D | k ) + k V ( C | k ) ,
where 0 k n 1 . Conditions (1) states that an individual always gains more from defecting than cooperating. Conditions (2) and (3) mean that an individual always gains more when there are more cooperators in the group. Conditions (4) indicates that the total payoff of the group is larger when there are more cooperators in the group.
We define a reactive strategy in the n -player repeated PD as follows: an individual with a reactive strategy cooperates with a certain probability in the first round, and from the second round on, cooperates with a certain probability determined by the number of opponents (i.e., group members other than the self) who have cooperated in the preceding round. Any reactive strategy is described by a vector d = ( d φ ,   d 0 ,   d 1 ,   ,   d n 1 ) , where d φ denotes the probability with which an individual obeying this strategy cooperates in the first move and d j denotes the probability with which he cooperates in a given round, provided that j of the n 1 opponents have cooperated in the preceding round. We focus on the set, Ω , of all reactive strategies in which d φ is either 0 or 1 and d j is either 0 or 1, namely, Ω = { ( d φ , d 0 ,   d 1 ,   ,   d n 1 ) | d φ { 0 ,   1 } ,   d j { 0 ,   1 } ,   j { 0 ,   1 , , n 1 } } . Obviously, there are 2 n + 1 strategies in Ω .
Table 1 shows the payoff matrix of the general two-strategy n -player game, where a l denotes the payoff to an individual with strategy A whose opponents are n l individuals with strategy A and l 1 individuals with strategy B; likewise, b l denotes the payoff to an individual with strategy B confronting with the same composition of opponents. Let A and B in Table 1 represent a strategy d in Ω and ALLD, respectively. To describe the game involving d and ALLD, we need to write down all a l and b l in Table 1 in terms of δ , V ( C | k ) , V ( D | k ) , and d . Consider a group consisting of k individuals adopting d and n k individuals adopting ALLD. We denote by h k the expected number of rounds in which a d individual cooperates before the group is dismissed. Since there are on average 1 / ( 1 δ ) rounds, we have 0 h k 1 / ( 1 δ ) . Further, since we do not consider any errors in behavior, d individuals perfectly coordinate their behavior, that is, either all or none of them cooperate in a given round. Hence, we obtain:
a n k + 1 = h k V ( C | k ) + ( 1 1 δ h k ) V ( D | 0 ) ,
b n k = h k V ( D | k ) + ( 1 1 δ h k ) V ( D | 0 ) .
For any group composition, h k is determined by d φ , d k 1 , and d 0 in the following manner. First, suppose that d φ = d k 1 = 1 . In this case, d never defects and, hence, h k = 1 / ( 1 δ ) . Second, if d φ = 1 and d k 1 = d 0 = 0 , d cooperates only in the first round and, thus, h k = 1 . Third, if d φ = d 0 = 1 and d k 1 = 0 , d alternates cooperation and defection indefinitely, in which case h k = 1 / ( 1 δ 2 ) . Fourth, if d φ = d 0 = 0 , d defects in every round and hence h k = 0 . Fifth, if d φ = 0 and d 0 = d k 1 = 1 , d defects only in the first round and thus h k = δ / ( 1 δ ) . Sixth, if d φ = d k 1 = 0 and d 0 = 1 , d alternates defection and cooperation indefinitely, in which case h k = δ / ( 1 δ 2 ) . In sum, h k is given by:
h k = d φ [ 1 + δ d k 1 1 δ + δ 2 d 0 ( 1 d k 1 ) 1 δ 2 ] + ( 1 d φ ) d 0 [ δ 1 δ 2 + d k 1 δ 2 1 δ 2 ] .
Therefore, an n -player repeated PD involving d and ALLD is specified using Equations (5)–(7).

2.2. Evolutionary Dynamics

We use a model of stochastic evolutionary dynamics developed for the general two-strategy n -player game [18]. A population consisting of i individuals adopting strategy A and N i individuals adopting strategy B is considered, where the population size, N , is constant ( 0 i N ) . Groups of n individuals are formed by choosing individuals at random from the population and a game is played within each of these groups ( 2 n N ) . The payoffs of the game are as shown in Table 1. Let F i and G i be the expected payoffs to an A individual and a B individual, respectively, when the number of A individuals in the population is i . The fitness of A and B individuals are given by f i = 1 w + w F i and g i = 1 w + w G i , respectively, where w is the selection intensity ( 0 < w < 1 ) . The population dynamics are formulated as a Moran process with frequency-dependent fitness: at each time step, an individual is chosen for reproduction with the probability proportional to its fitness, and then one identical offspring is produced to replace another individual randomly chosen for death with probability 1/N [17,51]. Denote by ρ A , B the probability that a single individual with strategy A in a population of N − 1 individuals with strategy B will finally take over the whole population (i.e., the fixation probability). Throughout the paper, we assume that the interaction group is smaller than the population ( N > n ) . Here, we have:
ρ A , B = 1 / ( 1 + l = 1 N 1 i = 1 l 1 w + w k = 1 n ( i n k ) ( N i 1 k 1 ) ( N 1 n 1 ) b k 1 w + w k = 1 n ( i 1 n k ) ( N i k 1 ) ( N 1 n 1 ) a k ) ,
where ( s t ) represents a binomial coefficient when s t and is defined zero when s < t .
Assuming that selection is weak ( w 1 ), [19,52,53] have shown that ρ A , B is given approximately by:
ρ A , B 1 N 1 1 α w / [ n ( n + 1 ) ] ,  
where:
α = ( N n ) k = 1 n k ( a k b k ) + ( n + 1 ) k = 1 n 1 k ( a k + 1 b k )
(see also [18,54,55]). Since ρ A ,   B increases with α , the strategy with the largest fixation probability maximizes the right-hand side of Equation (10).

3. Results

3.1. General Case

We define the best reactive strategy as the strategy (or strategies) that maximizes the fixation probability when introduced as a single mutant in a population of ALLD. The best reactive strategy maximizes the right-hand side of Equation (10) in the context of the n -player repeated PD specified by Equations (5)–(7). Equation (10) is equivalent to:
α = ( N n ) ( k = 1 n a n k + 1 n b n ) + ( N + 1 ) k = 1 n ( n k ) ( a n k + 1 b n k ) .
Putting Equations (5) and (6) into Equation (11) and explicitly denoting the dependence on strategy d Ω , we obtain:
α ( d ) = k = 1 n h k ( d ) f ( k ) ,
where:
f ( k ) = ( N n ) ( V ( C | k ) V ( D | 0 ) ) + ( N + 1 ) ( n k ) ( V ( C | k ) V ( D | k ) ) ,
is independent of strategy d . Note that h k ( d ) is given by Equation (7) and is dependent on d in the following manner:
h k ( d ) = h ( d φ ,   d 0 , d k 1 , k ) .
The best reactive strategy d is given by the following equation:
d = argmax d Ω ( k = 1 n h k ( d ) f ( k ) ) .
As shown in Appendix A, the solution of the maximization problem in Equation (15) is given as follows. On the one hand, when:
1 1 δ k = 1 n max { f ( k ) ,   0 } + k = 1 n min { f ( k ) ,   0 } > 0 ,
the solution of Equation (15) is given by:
{ d φ = 1 , d k 1 = 1 if   f ( k ) > 0 , d k 1 = 0 if   f ( k ) < 0 ,  
for 1 k n . Note that d k 1 can be either 0 or 1 in case f ( k ) = 0 . On the other hand, when the inequality in Inequality (16) is reversed, the solution of Equation (15) is given by:
d φ = d 0 = 0 ,
where d k 1 can be either 0 or 1 for 2 k n . In sum, Conditions (16)–(18) specify the best reactive strategy, d .
Our main result Conditions (16)–(18) states that so long as the game is repeated sufficiently many times (i.e., δ is sufficiently large), each bit, d k 1 , of the best reactive strategy is solely determined by the sign of f ( k ) , which is determined by the payoff structure of the game (and the group size and the population size. In Equation (13), the term V ( C | k ) V ( D | 0 ) is the payoff to a cooperating player in a group where k players cooperate minus the payoff to a defecting player in a group where no one cooperates. Thus, this term represents the benefit of cooperation. The term V ( C | k ) V ( D | k ) represents the payoff difference between a cooperating player and a defecting player in a group where k individuals cooperate. Thus, this negative term represents the cost of cooperation. Weighted sum of them determines whether the best reactive strategy should cooperate if k 1 players have cooperated in the previous round.
In the following analysis, we consider cases in which Inequality (16) holds unless otherwise stated.

3.2. Linear Payoff

Let us begin with a special case where the following conventional linear payoff assumption holds:
V ( D | k ) = b k n ,
V ( C | k ) = V ( D | k ) c .
From Inequalities (1)–(4), it has to be that 0 < c < b < n c . Putting Equations (19) and (20) into Equation (13), we obtain:
f ( k ) = ( N n ) ( b k n c ) ( N + 1 ) ( n k ) c .
Since f ( k ) increases linearly with k in this case, there exists a critical value above which f ( k ) is positive (Figure 1a. Hence, from Equation (17), α is maximized when the following condition is met:
{ d k = 1   for   k > k , d k = 0   for   k < k ,
where:
k = ( n + 1 ) N ( N n ) b / ( n c ) + N + 1 1 .
From Equation (23), k decreases with increasing N , and for an infinitely large population, k is given by:
k = n + 1 b / ( n c ) + 1 1 .
Since b / ( n c ) < 1 by assumption, k > ( n 1 ) / 2 holds. Therefore, any strategy d that cooperates when a half or less of the opponents have cooperated in the previous round cannot be the best reactive strategy.
A similar, but slightly different, result has been obtained by Kurokawa et al. (2010) [19]. They examined the fixation probability of a strategy called TFTa, being introduced as a mutant into a population of ALLD, under the linear payoff assumption as given by Equations (19) and (20). An individual adopting TFTa cooperates in the first round, and from the second round on, cooperates if and only if a or more of the opponents have cooperated in the previous round, where a { 0 ,   1 ,   n 1 } measures the level of generosity [23]. Thus, there are n strategies that are classified as TFTa with different levels of generosity. From these, Kurokawa et al. (2010) specified a o p t , the optimal level of generosity, which maximizes the fixation probability. In the terminology of the present study, the n strategies constitute the subset of Ω for which d φ = 1 and d 0 d 1 d n 1 = 1 . In addition to these, we consider two kinds of non-TFTa reactive strategies. One is those that defect in the first round (i.e., d φ = 0 . Our analysis shows that some of this kind can maximize the fixation probability if the expected number of rounds is sufficiently small, partially modifying Kurokawa et al.’s (2010) finding. Another kind is what we call “paradoxical” strategies (see Introduction), in which d k > d k + 1 holds for at least some k ; for example, d = ( 1 ,   0 ,   1 ,   1 ,   0 ,   ,   0 ,   1 ) . It turns out that Equations (22)–(24) is equivalent to the optimal level of generosity obtained by Kurokawa et al. (2010). Thus, their result is not changed by taking paradoxical strategies into consideration.

3.3. Non-Linear Payoff

The n-player PD can be further subcategorized according to characteristics of the fitness functions [39,46]. To illustrate, we define the cost and benefit associated with cooperation as:
c ( k ) = V ( D | k ) V ( C | k ) ,  
b ( k ) = V ( D | k ) ,
respectively. From Inequalities (1) and (2), c ( k ) > 0 is required for all k . We can measure how the cost and benefit depend on the number of cooperators by considering:
Δ c ( k ) = c ( k + 1 ) c ( k ) ,  
Δ b ( k ) = b ( k + 1 ) b ( k ) .  
For example, when Δ c ( k ) = 0 holds for all k , the cost is constant. Let us also consider:
Δ 2 b ( k ) = Δ b ( k + 1 ) Δ b ( k ) ,  
for 0 k n 2 . It should be intuitively clear that the n-player PD requires Δ b ( k ) > 0 , which is equivalent to (3), while Δ 2 b ( k ) can be either positive or negative (or zero. If, for instance, Δ 2 b ( k ) 0 holds for all k , the benefit increases in an accelerative manner with the number of cooperators (e.g., the weakest-link game; [56]), whereas if Δ 2 b ( k ) 0 always holds, the increase is decelerated (e.g., the volunteer’s dilemma; [57]. The class of n-player PDs for which Δ 2 b ( k ) = 0 always holds represent the public goods games in which the amount of public goods increases linearly with the number of cooperators [47].
The linear payoff assumptions in Equations (19) and (20) mean that the following conditions are satisfied for all k :
Δ 2 b ( k ) = 0 ,
Δ c ( k ) = 0 .
Below we examine how evolution of generosity and paradoxical behavior may be affected by relaxing Equations (30) and (31).

3.3.1. Evolution of Generosity with Non-Linear Payoff

Constant Cost with Non-Linear Benefit

Consider the case when there is no restriction on the benefit function (except Inequalities (1)–(4)), that is, Equation (30) is not necessarily true. For the moment, we assume Equation (31). From Equations (13) and (20), we have:
f ( k ) = ( N n ) ( V ( D | k ) V ( D | 0 ) ) [ ( n + 1 ) N k ( N + 1 ) ] c .  
Using Conditions (1) and (20), we obtain:
V ( D | k ) V ( D | 0 ) = l = 1 k [ V ( D | l ) V ( C | l ) + V ( C | l ) V ( D | l 1 ) ] < k c ,
which gives:
f ( k ) < [ ( 2 N n + 1 ) k ( n + 1 ) N ] c .
Thus, we obtain a necessary condition for f ( k ) > 0 , namely:
k > ( n + 1 ) N 2 N n + 1 .  
The right-hand side of Inequality (35) decreases and approaches ( n + 1 ) / 2 with increasing N . Hence, from Equation (17), d k = 1 holds in the best reactive strategy only if k > ( n 1 ) / 2 . This means that any strategy d who cooperates when a half or less of the opponents have cooperated in the previous round cannot be the best reactive strategy.

Variable Cost with Non-Decelerating Benefit

Let us now remove the restriction on the cost function, that is, we consider the case when Equation (31) does not necessarily hold. Instead of Equation (30), we assume:
Δ 2 b ( k ) 0 ,
which means that the benefit from cooperation increases with the number of cooperators in either linear or accelerative fashion. From Equations (13), (26) and (28), we have:
f ( k ) = ( N n ) ( l = 0 k 1 Δ b ( l ) ) [ ( N + 1 ) ( n k ) + ( N n ) ] c ( k ) .
Using Inequality (36), we obtain:
l = 0 k 1 Δ b ( l ) k Δ b ( k 1 ) .
From Conditions (37) and (38), we have:
f ( k ) ( N n ) k Δ b ( k 1 ) [ ( N + 1 ) ( n k ) + ( N n ) ] c ( k ) .
Meanwhile, using Conditions (1), (25), (26) and (28), we obtain:
Δ b ( k 1 ) = V ( D | k ) V ( D | k 1 ) < V ( D | k ) V ( C | k ) = c ( k ) .
Therefore, combining Equation (39) and (40), we obtain:
f ( k ) < [ ( 2 N n + 1 ) k ( n + 1 ) N ] c ( k ) .
Hence, as in the case of constant cost, Inequality (35) is necessary for f ( k ) > 0 , which again means that the best reactive strategy must defect when a half or more of the opponents have defected in the previous round.

A Numerical Example

The results for the above two special cases are consistent with our finding in the analysis under the linear payoff assumption. However, the result may be altered qualitatively in the absence of any restriction on the benefit or cost functions. As an illustration, consider the following payoff functions:
V ( D | k ) = 1 ( n k + 1 n + 1 ) 2 ,
V ( C | k ) = 1 ( n k + 2 n + 1 ) 2 1 2 ( n + 1 ) 2 ,  
which satisfies Inequalities (1)–(4), but is not consistent with Equations (30) or (31). To be specific, let us set N = 400 , n = 30 , and δ = 0.8 . Putting them into Equation (13), we obtain:
f ( k ) = 25400 31 + 146407 k 1922 1172 k 2 961 .
In this case, Inequality (16) is met, and as shown in Figure 1b, f ( k ) > 0 holds if and only if k 14 . Hence, from Equation (17), the best reactive strategy ( d ) satisfies the following:
{ d φ = 1 , d k = 1 for   k 13 , d k = 0 for   k 12 .
Remember that d k is defined as the probability with which he cooperates in a given round, provided that k of the n 1 “opponents” have cooperated in the preceding round. This demonstrates that there exist situations in which such a strategy that cooperates when a half or more of the opponents have defected in the previous round can be the best reactive strategy.
For comparison, the second and third best strategies following Equation (45) are given by:
{ d φ = 1 , d k = 1 for   k 14 , d k = 0 for   k 13 ,
{ d φ = 1 , d k = 1 for   k 12 , d k = 0 for   k 11 ,
respectively. For w = 0.001 , the approximated fixation probabilities for Equations (45)–(47), calculated from Equation (9), are 0.00254215, 0.00254207, and 0.00254176, respectively.

3.3.2. Evolution of Paradoxical Behavior with Non-Linear Payoff

Non-Increasing Cost and Non-Linear Benefit

Consider the case when there is no restriction on the benefit function (except Inequalities (1)–(4)) and, thus, Equation (30) is not necessarily true. As for the cost function, we partially relax Equation (31) by assuming instead:
Δ c ( k ) 0 ,
which means that the cost of cooperation either is constant or decreases with increasing number of cooperators. Equation (13) changes into:
f ( k ) = ( N n ) ( V ( C | k ) V ( D | 0 ) ) ( N + 1 ) ( n k ) c ( k ) .  
Hence, we have:
f ( k + 1 ) = ( N n ) ( V ( C | k + 1 ) V ( D | 0 ) ) ( N + 1 ) ( n k 1 ) c ( k + 1 ) .
From Conditions (48)–(50), we have:
f ( k + 1 ) > f ( k ) .
This means that if f ( k ) > 0 then f ( m ) > 0 for any m > k . Hence, from Equation (17), a paradoxical strategy cannot be the best reactive strategy.

A Numerical Example

Our analysis thus far has shown that a paradoxical strategy cannot be the best reactive strategy so long as the cost of cooperation is either unaltered or decreased by increasing number of cooperators. However, this may not be the case if the cost of cooperation sometimes increases with the number of cooperators. Let us provide a numerical example in which the best reactive strategy exhibits a paradoxical behavior. We set N = 400 , n = 30 , δ = 0.8 , and b = 1 and assume that Equation (30) is met. Instead of Equation (31), here we assume:
c ( k ) = { 0.07   for   k 20 , 0.04   for   k 19 ,  
for which Inequalities (1)–(4) are satisfied. This represents a case when the cost of cooperation is larger in the presence of more cooperators in the group. Intuitively, in such a situation, a paradoxical strategy that cooperates only when there are relatively few cooperators may be more advantageous than non-paradoxical strategies that, which cooperate whenever the number of cooperators exceeds a threshold. Indeed, since in this case Inequality (16) is met, and f ( k ) is non-monotonic in this case (Figure 1c), the best reactive strategy ( d ) is given from Equation (17) by:
{ d φ = 1 , d k = 1 for   k = 17 ,   18 ,   k 21 , d k = 0 for   k 16 ,   k = 19 ,   20 .  
That is, the best reactive strategy cooperates when there are 17 or more opponents who have cooperated in the previous round, except when the number of cooperators is 19 or 20, in which case it defects. The relative advantages of cooperation and defection are determined by a balance between the following two factors: On the one hand, obviously, cooperation is more advantageous when the associated cost is smaller. On the other hand, given that the cost is constant, the relative advantage of cooperation increases with the number of cooperators in the group. In the current example, there is a rise of the cost at k = 20 (see Equation (52)), which can be compensated for only when the number of cooperating opponents increases to 21 or more.
The second and third best strategies following Equation (53) are given respectively by:
{ d φ = 1 , d k = 1 for   k = 16 ,   17 ,   18 ,   k 21 , d k = 0 for   k 15 ,   k = 19 ,   20 ,
{ d φ = 1 , d k = 1 for   k = 18 ,   k 21 , d k = 0 for   k 17 ,   k = 19 ,   20 .
For w = 0.001 , the fixation probabilities of strategies Equations (53)–(55) with the small w approximation are 0.00251170, 0.00251155, and 0.00251154, respectively.

3.4. The Best Reactive Strategy under Moderate Selection Intensity

Thus far we have assumed that selection is sufficiently weak ( w 1 ) so that the fixation probabilities can be approximated by Equation (9). Here we numerically obtain the exact fixation probabilities using (8) without the weak selection assumption to examine how this may affect the identity of the best reactive strategy.
For simplicity, we make the linear payoff assumption, Equations (19) and (20). Figure 2 illustrates how the best reactive strategy ( d ), obtained by using Equation (8), changes depending on the values of δ and w for parameter values b = 9 , c = 1 , N = 50 , and n = 10 . For this parameter setting, a comparison of fixation probabilities calculated from Equation (9) would indicate TFT6 as the best reactive strategy when δ > δ holds, where δ 0.49 , and d φ = d 0 = 0 when δ < δ . This gives a good approximation of the exact numerical solution for w = 0.001 as shown in Figure 2. However, when w is larger, the approximation is no longer valid. Figure 2 shows that there exist situations in which such a strategy that cooperates when a half or more of the opponents have defected in the previous round can be the best reactive strategy, which cannot be the case in the limit of week selection. A reactive strategy’s fixation probability is determined by its payoffs relative to the unconditional defector at various frequencies ( i   = 1 , 2 , , N 1 in Equation (8). Among the N 1 values, a relative payoff at a lower frequency has a greater impact on the fixation probability. Hence, if, as numerically suggested by Kurokawa et al. (2010), the relative payoff at a low frequency tends to be larger for more generous reactive strategies, this is likely to favor the evolution of generosity. This effect might be more pronounced when selection, rather than random drift, plays a greater role. These considerations all in all accord with our numerical example suggesting that more generous reactive strategies are selectively favored when selection is more intense. It was also numerically shown that δ can affect the identity of the best reactive strategy, while in the limit of weak selection δ does not affect what the best reactive strategy is as far as it satisfies Inequality (16). In addition, within this particular range of parameter values, we did not find any case in which a paradoxical strategy is the best reactive strategy.

4. Discussion

We have investigated stochastic evolutionary dynamics of a population in which an individual’s fitness is determined by the n-player repeated PD. We have compared the fixation probabilities of different reactive strategies when they appear as a rare mutant in a population of unconditional defectors. Reactive strategies in our analysis are described by a vector d = ( d φ ,   d 0 ,   d 1 ,   ,   d n 1 ) , where d φ represents the probability with which the strategy cooperates in the first round and d j represents the probability with which the strategy cooperates in a given round of a repeated game when j of the n 1 opponents have cooperated in the previous round. We have considered a set, Ω , of all reactive strategies for which d φ is either 0 or 1 and d j is either 0 or 1 to specify the best reactive strategy attaining the maximum fixation probability. In a repeated PD, after one round is played, another round will be played with probability δ . Under the assumption of weak selection, we have specified a threshold of δ below which all strategies satisfying d φ = d 0 = 0 are the best reactive strategies. We have also found that when δ exceeds the threshold, d k 1 of the best reactive strategy is solely determined by the sign of f ( k ) , as given by Equation (13), which is interpreted as a weighted sum of the cost and benefit of cooperation when k 1 opponents have cooperated in the proceeding round.
The present study extends our previous analysis (Kurokawa et al. (2010), [19]) by considering a broader set of reactive strategies and non-linear payoff functions. We also investigate how robust the indicated identity of the best reactive strategy is to deviations from the weak selection assumption. Kurokawa et al. (2010) specified the best reactive strategy from n reactive strategies classified as TFTa, which constitute a subset of Ω , under the conventional linear payoff assumption. One of their findings was that any TFTa that tolerates defection by more than a half of the group members can never be the best reactive strategy. The present study has shown, on the one hand, that under the linear payoff assumption, the previous finding holds true even for all reactive strategies in Ω . For non-linear payoff functions, on the other hand, we have shown that the previous finding does not always hold, that is, when the benefit increases in a decelerative manner with the number of cooperators, the best reactive strategy tolerates defection by more than a half of the opponents. We have also found that a strategy tolerating defection by more than a half of the opponents can be the best reactive strategy when selection is not weak.
We have also demonstrated that a paradoxical strategy, in which d k > d k + 1 holds for at least some k , can be the best reactive strategy when the cost of cooperation increases with the number of cooperators in a group. As far as we know, it has not been pointed out that a conditional cooperator can sometimes do better by behaving paradoxically in competition with defectors. A potentially relevant observation is that human participants in laboratory experiments of the repeated PD sometimes behave as if they were following a paradoxical strategy (e.g., [58,59,60,61]), though these studies assume linear payoff functions. The present study suggests that it may be of interest to investigate the effect of the shape of the payoff functions on the occurrence of paradoxical behaviors in the laboratory setting.
So far, we have examined the fixation probability of strategy d when appeared as a single mutant in a population of ALLD. In Appendix B we also compare the fixation probability of ALLD when introduced as a single mutant in a population of reactive strategy d , across all d in Ω . A reactive strategy associated with a lower fixation probability of ALLD is regarded as more robust against invasion by ALLD. The reactive strategy that minimizes the fixation probability of ALLD turns out to be TFTn-1, that is, d = ( 1 , 0 ,   0 ,   ,   0 ,   1 ) , when δ is larger than the threshold specified by Inequality (16), and the strategies satisfying d φ = d 0 = 0 when otherwise. Further, we investigate the ratio of the fixation probability of d in a population of ALLD to the fixation probability of ALLD in a population of d , across all d in Ω . As shown in Appendix C, the reactive strategy maximizing the ratio of the fixation probabilities is TFTn-1 when δ is larger than the threshold, and d φ = d 0 = 0 when otherwise. These results are consistent with our earlier analysis [19], even though the present study considers paradoxical strategies and non-linear payoff functions, which were not considered in the previous study.
Natural selection is regarded as favoring a mutant reactive strategy replacing a population of ALLD if the fixation probability of the reactive strategy exceeds 1 / N , which is the fixation probability for neutral evolution. The reactive strategies that are the best when δ is small so that Inequality (16) is not satisfied (i.e., d φ = d 0 = 0 ) always have the fixation probability 1 / N . Thus, when Inequality (16) is not met, the best reactive strategies and ALLD are selectively neutral. On the other hand, when δ is large so that Inequality (16) is satisfied, there are other reactive strategies, including the best one, with the fixation probability larger than 1 / N . Hence, in this case, the best reactive strategy is always selectively favored over ALLD. See Appendix D for a further analysis on the case assuming linear payoff functions.
Thus far, our focus has been on the best reactive strategy, which maximizes the fixation probability. However, it is also of interest to examine characteristics of other reactive strategies whose fixation probabilities are relatively high and difference in fixation probabilities among those best strategies. For example, Table 2 gives a list of ten best strategies in the case of n = 10 assuming Equations (19) and (20). Figure 3 illustrates the fixation probabilities of the 32 possible strategies in the case of n = 4 for various values of δ , under the assumption of Equations (19) and (20). The value of δ makes difference in the order of the strategies. Note that as shown in Appendix A, the value of δ does not have an effect on the order of the strategies belonging to subset of Ω d φ , d 0 .
Finally, let us conclude by mentioning two issues to be addressed in future research. First, actual human behavior can be more complex than that in our model. Although the present study considers a greater variety of reactive strategies than in a previous study [19], the strategy set can be further broadened. Our strategy set, Ω , does not include, for example, continuous strategies that engage in partial cooperation (see [62]), or those that make decisions referring to their own past behaviors ([63,64,65,66,67,68]. In addition, in our analysis, the benefit, b , and the cost, c , of cooperation are assumed to be the same for all individuals. Namely, we assumed that there are no differences among agents in this sense. In reality, however, individuals may have different values of b and c , and these variations could affect the model outcome [69]. Secondly, we have assumed that perfect information regarding group members’ behaviors is always available. It has been shown, however, that results of related models may be qualitatively altered depending on the accessibility of the relevant information [45,64,70,71,72,73,74,75].

Author Contributions

Every author was involved in all stages of development and writing of the paper.

Funding

S.K. is partially funded by Chinese Academy of Sciences President’s International Fellowship Initiative. Grant No. 2016PB018 and Monbukagakusho grant 16H06412. J.Y.W. is partly supported by JSPS Kakenhi 16K05283 and 16H06412. Y.I. is partly supported by the JSPS aid for Topic-Setting Program to Advance Cutting-Edge Humanities and Social Sciences Research and Kakenhi 16K07510 and 17H06381.

Acknowledgments

We thank Yoshio Kamijo (Kochi University of Technology) for discussions.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Solving (15)

To solve the maximization problem Equation (15), we first consider subsets Ω u ,   v of reactive strategies satisfying d φ = u and d 0 = v , where:
Ω 1 , 0 Ω 1 , 1 Ω 0 , 0 Ω 0 , 1 = Ω ,
we derive the solution, d k 1 ( 2 k n ), within each subset, after which the solution in Ω is obtained by comparing the maxima of α within the subsets.
First, when d φ = 1 and d 0 = 0 , Equation (7) gives:
h k ( d ) = 1 + δ d k 1 1 δ .  
Note that in this case h k ( d ) depends only on a single component ( d k 1 ) of d . Since f ( k ) is constant and the right-hand side of Equation (A2) increases with d k 1 , the solution of the maximization problem within subset Ω 1 , 0 is given by:
{ d k 1 = 1   if   f ( k ) > 0 , d k 1 = 0   if   f ( k ) < 0 ,  
for 2 k n . In case f ( k ) = 0 , both d k 1 = 1 and d k 1 = 0 are the solutions. Hence, from Equations (12), (A2) and (A3), we obtain the maximum of α within Ω 1 , 0 as follows:
α ( d ) = 1 1 δ k = 1 n max { f ( k ) ,   0 } + k = 1 n min { f ( k ) ,   0 } .  
Note that we use f ( 1 ) < 0 (from Conditions (1), (2) and (13)) in the derivation of Equation (A4).
Second, when d φ = d 0 = 1 , we have:
h k ( d ) = 1 + δ d k 1 1 δ 2 ,  
which increases with d k 1 . Thus, the solution in this case is given as well by Equation (A3) for 2 k n . Accordingly, the maximum of α within Ω 1 , 1 is given by:
α ( d ) = 1 1 δ k = 1 n max { f ( k ) ,   0 } + 1 1 δ 2 k = 1 n min { f ( k ) ,   0 } + δ 1 δ 2 f ( 1 ) .  
Third, when d φ = d 0 = 0 , we have:
h k ( d ) = 0 .  
This means that all combinations of d k 1 for 2 k n are the solutions and the maximum of α within Ω 0 , 0 is:
α ( d ) = 0 .
Fourth, when d φ = 0 and d 0 = 1 , we obtain:
h k ( d ) = δ ( 1 + δ d k 1 ) 1 δ 2 ,
which increases with d k 1 . Hence, the solution in this case is also given by Equation (A3) for 2 k n . Therefore, the maximum of α within Ω 0 , 1 is given by:
α ( d ) = δ 1 δ k = 1 n max { f ( k ) ,   0 } + δ 1 δ 2 k = 1 n min { f ( k ) ,   0 } + δ 2 1 δ 2 f ( 1 ) .
Finally, a comparison of Equations (A4), (A6), (A8) and (A10) shows that the solution of Equation (15) depends on the expected number of rounds, 1 / ( 1 δ ) , as shown in Conditions (16)–(18). Note that as Equations (A2), (A5), (A7) and (A9) indicate, δ does not affect the rank order of reactive strategies in the magnitudes of the fixation probabilities within each subset.

Appendix B

The Fixation Probability of ALLD as a Single Mutant in a Population of d

Under the assumption of weak selection, the fixation probability, ρ B ,   A , of a single individual with strategy B in a population of N 1 individuals with strategy A is given approximately by:
ρ B , A 1 N 1 1 φ w / [ n ( n + 1 ) ] ,
where:
φ = ( N n ) ( n a 1 k = 1 n b k ) ( N + 1 ) k = 1 n ( n k ) ( a k + 1 b k ) .
Let us specify the reactive strategy d that minimizes the fixation probability of ALLD. Putting Equations (5) and (6) into Equation (B2), we obtain:
φ ( d ) = k = 1 n 1 h k ( d ) g ( k ) ( N n ) n h n ( d ) ( V ( C | n ) V ( D | 0 ) ) ,
where:
g ( k ) = [ ( N n ) ( V ( D | k ) V ( D | 0 ) ) + ( N + 1 ) k ( V ( D | k ) V ( C | k ) ) ] ,
is independent of strategy d . Since V ( D | k ) V ( D | 0 ) and V ( D | k ) > V ( C | k ) always hold, we have g ( k ) > 0 .
Firstly, when d φ = 1 and d 0 = 0 , Equations (7) and (A13) show:
φ = k = 1 n 1 [ 1 + δ d k 1 1 δ ] g ( k ) ( N n ) n [ 1 + δ d n 1 1 δ ] ( V ( C | n ) V ( D | 0 ) ) .
From Equation (A15), since g ( k ) > 0 and V ( C | n ) > V ( D | 0 ) always hold, φ is minimized when d φ = 1 , d k = 0 for 0 k n 2 and d n 1 = 1 . This means that the fixation probability of ALLD is minimized in a population of strategy d = ( 1 , 0 ,   0 ,   ,   0 , 1 ) , namely, TFTn−1 in Ω 1 , 0 . In this case, the minimum of φ is given by:
φ = k = 1 n 1 g ( k ) ( N n ) n 1 1 δ ( V ( C | n ) V ( D | 0 ) ) .
Secondly, when d φ = d 0 = 1 , (A13) becomes:
φ = k = 1 n 1 [ 1 1 δ 2 + δ d k 1 1 δ 2 ] g ( k ) ( N n ) n [ 1 1 δ 2 + δ d n 1 1 δ 2 ] ( V ( C | n ) V ( D | 0 ) ) .
Since g ( k ) > 0 and V ( C | n ) > V ( D | 0 ) , φ is minimized when d φ = 1 , d k = 0 for 1 k n 2 and d 0 = d n 1 = 1 . Hence, the minimum of φ in Ω 1 , 1 is given by:
φ = 1 1 δ [ ( N n ) ( V ( D | 1 ) V ( D | 0 ) ) + ( N + 1 ) ( V ( D | 1 ) V ( C | 1 ) ) ] + 1 1 δ 2 k = 2 n 1 g ( k ) ( N n ) n 1 1 δ ( V ( C | n ) V ( D | 0 ) ) .
Thirdly, when d φ = d 0 = 0 , Equations (7) and (A13) indicates that φ = 0 always holds. Hence, irrespective of the values of d k for 1 k n 1 , the minimum of φ in Ω 0 , 0 is:
φ = 0 ,
and fourthly, when d φ = 0 and d 0 = 1 , Equation (A13) becomes:
φ = k = 1 n 1 [ δ 1 δ 2 + d k 1 δ 2 1 δ 2 ] g ( k ) ( N n ) n [ δ 1 δ 2 + d n 1 δ 2 1 δ 2 ] ( V ( C | n ) V ( D | 0 ) ) .
Since g ( k ) > 0 and V ( C | n ) > V ( D | 0 ) , φ is minimized when d φ = 0 , d k = 0 for 1 k n 2 and d 0 = d n 1 = 1 . Thus, the minimum of φ in Ω 0 , 1 is:
φ = δ 1 δ [ ( N n ) ( V ( D | 1 ) V ( D | 0 ) ) + ( N + 1 ) ( V ( D | 1 ) V ( C | 1 ) ) ] + δ 1 δ 2 k = 2 n 1 g ( k ) ( N n ) n δ 1 δ ( V ( C | n ) V ( D | 0 ) ) .
A comparison of Equations (A16), (A18), (A19) and (A21) shows the following. The strategy that minimizes φ among all strategies in Ω is TFTn−1 when:
δ > 1 ( N n ) n ( V ( C | n ) V ( D | 0 ) ) k = 1 n 1 g ( k ) ,  
and any strategy satisfying d φ = d 0 = 0 when the inequality in Equation (A22) is reversed. This means that the fixation probability of ALLD is minimized in a population of strategy d = ( 1 ,   0 ,   0 ,   ,   0 , 1 ) , namely, TFTn-1 when δ exceeds a threshold and otherwise, it is minimized in a population of strategies satisfying d φ = d 0 = 0 .

Appendix C

The Ratio of the Fixation Probabilities

Ratio ρ A ,   B / ρ B ,   A can be viewed as a measure of the selective advantage of A over B for the long-term evolutionary process. From Equations (9) and (A11), the ratio is given approximately by:
ρ A ,   B ρ B ,   A 1 + σ w n ,
where:
σ = ( N n ) ( a 1 b n ) + N k = 1 n 1 ( a k + 1 b k ) .
We regard A and B as strategy d and ALLD, respectively, to specify the reactive strategy that maximizes the ratio of the fixation probabilities. Putting Equations (5) and (6) into Equation (A24) yields:
σ = N k = 1 n 1 h k ( V ( D | k ) V ( C | k ) ) + ( N n ) h n ( V ( C | n ) V ( D | 0 ) ) .  
Firstly, when d φ = 1 and d 0 = 0 , by using Equation (7), Equation (A25) becomes:
σ = N k = 1 n 1 [ 1 + δ d k 1 1 δ ] ( V ( D | k ) V ( C | k ) ) + ( N n ) [ 1 + δ d n 1 1 δ ] ( V ( C | n ) V ( D | 0 ) ) .  
Since V ( D | k ) V ( C | k ) > 0 and V ( C | n ) > V ( D | 0 ) always hold, σ is maximized when d φ = 1 , d k = 0 for 0 k n 2 and d n 1 = 1 . This means that the largest ratio of the fixation probabilities is attained by strategy d = ( 1 , 0 ,   0 ,   ,   0 , 1 ) , namely, TFTn−1. Hence, the maximum of σ in Ω 1 , 0 is given by:
σ = N k = 1 n 1 ( V ( D | k ) V ( C | k ) ) + ( N n ) 1 1 δ ( V ( C | n ) V ( D | 0 ) ) .
Secondly, when d φ = d 0 = 1 , Equation (A25) becomes:
σ = N k = 1 n 1 [ 1 1 δ 2 + δ 1 δ 2 d k 1 ] ( V ( D | k ) V ( C | k ) ) + ( N n ) [ 1 1 δ 2 + δ 1 δ 2 d n 1 ] ( V ( C | n ) V ( D | 0 ) ) .
Since V ( D | k ) V ( C | k ) > 0 and V ( C | n ) > V ( D | 0 ) , σ is maximized when d φ = 1 , d k = 0 for 1 k n 2 and d 0 = d n 1 = 1 . Thus, the maximum of σ in Ω 1 , 1 is:
σ = N [ 1 1 δ ] ( V ( D | 1 ) V ( C | 1 ) ) N k = 2 n 1 [ 1 1 δ 2 ] ( V ( D | k ) V ( C | k ) ) + ( N n ) [ 1 1 δ ] ( V ( C | n ) V ( D | 0 ) ) .  
Thirdly, when d φ = d 0 = 0 holds, Equation (7) and Equation (A25) indicate that σ = 0 is always true, irrespective of the values of d k for 1 k n 1 . Therefore, the maximum of σ in Ω 0 , 0 is:
σ = 0 .
Fourthly, when d φ = 0 and d 0 = 1 , by using Equation (7), Equation (A25) becomes:
σ = N k = 1 n 1 [ δ 1 δ 2 + d k 1 δ 2 1 δ 2 ] ( V ( D | k ) V ( C | k ) ) + ( N n ) [ δ 1 δ 2 + d n 1 δ 2 1 δ 2 ] ( V ( C | n ) V ( D | 0 ) ) .
Since V ( D | k ) V ( C | k ) > 0 and V ( C | n ) > V ( D | 0 ) always hold, σ is maximized when d φ = 0 , d k = 0 for 1 k n 2 and d 0 = d n 1 = 1 . Hence, the maximum of σ in Ω 0 , 1 is given by:
σ = N [ δ 1 δ ] ( V ( D | 1 ) V ( C | 1 ) ) N k = 2 n 1 [ δ 1 δ 2 ] ( V ( D | k ) V ( C | k ) ) + ( N n ) [ δ 1 δ ] ( V ( C | n ) V ( D | 0 ) ) .
Finally, by comparing Equations (A27), (A29), (A30) and (A32), we obtain the following result. Of all strategies in Ω , the strategy maximizing σ is TFTn−1 when:
δ > 1 ( N n ) ( V ( C | n ) V ( D | 0 ) ) N k = 1 n 1 ( V ( D | k ) V ( C | k ) ) ,
and any strategy satisfying d φ = d 0 = 0 when the inequality in Inequality (A33) is reversed. This means that the largest ratio of the fixation probabilities is attained by strategy d = ( 1 ,   0 ,   0 ,   ,   0 , 1 ) , namely, TFTn-1 when Inequality (A33) is met and, otherwise, the largest ratio of the fixation probabilities is attained by strategies satisfying d φ = d 0 = 0 .

Appendix D

The Conditions for Reactive Strategies to Be Selectively Favored over ALLD

We investigate whether natural selection favors strategy d over ALLD under the assumption that the payoffs are given by linear functions in Equations (19) and (20). First, let us specify the condition for d replacing a population of ALLD to be selectively favored, which is true when the fixation probability of d exceeds 1 / N . Unless d φ = d 0 = 0 holds true, this condition is given by:
b n c > k = 0 n 1 d k ( n k ) k = 0 n 1 ( n k ) + d 0 ( 1 1 δ ) k = 0 n 1 d k ( k + 1 ) k = 0 n 1 ( k + 1 ) + d 0 ( 1 1 δ ) .
To make a geometric representation of the condition, let strategy d be represented by a point, Pd, on the x y -plane, whose coordinates are given by:
( x , y ) = ( k = 0 n 1 d k ( k + 1 ) k = 0 n 1 ( k + 1 ) + d 0 ,   k = 0 n 1 d k ( n k ) k = 0 n 1 ( n k ) + d 0 ) .
This formulation makes the right-hand side of Equation (A23) equivalent to the slope of the line through points Pd to M, the latter of which are given by x = y = 1 1 / δ . Then, Inequality (A34) is satisfied if and only if Pd is located below the “critical line,” which passes through M with slope b / ( n c ) , where 1 / n < b / ( n c ) < 1 . Figure A1 provides such a geometric representation for the case of n = 3 , showing d = ( 1 ,   0 ,   0 ,   1 ) and ( 1 ,   0 ,   1 ,   1 ) , but none of the others are selectively favored when replacing a population of ALLD. If d φ = d 0 = 0 holds true, the fixation probability of d in a population of ALLDs is equal to 1 / N .
Second, we specify the condition for ALLD replacing a population of d to be selectively disfavored, or the condition under which the fixation probability of ALLD is smaller than 1 / N . When either d φ > 0 or d 0 > 0 is satisfied, the condition is given by:
b n c > k = 0 n 1 d k ( k + 1 ) k = 0 n 1 ( k + 1 ) + d 0 ( 1 1 δ ) 2 d n 1 k = 0 n 1 d k ( k + 1 ) k = 0 n 1 ( k + 1 ) + d 0 ( 1 1 δ ) .
Let strategy d be represented by point Qd, whose coordinates are:
( x , y ) = ( 2 d n 1 k = 0 n 1 d k ( k + 1 ) k = 0 n 1 ( k + 1 ) + d 0 , k = 0 n 1 d k ( k + 1 ) k = 0 n 1 ( k + 1 ) + d 0 ) .
Figure A1. Geometric representations of Inequalities (A34) and (A36) for n = 3 . Each reactive strategy d = ( d φ , d 0 ,   d 1 ,   d 2 ) is represented by two points on the x y -plane, P( d φ , d 0 ,   d 1 ,   d 2 ) and Q( d φ , d 0 ,   d 1 ,   d 2 ). The empty square represents point M, which is given by x = y = 1 1 / δ and on the line y = x ( x < 0 ) (the broken line). The thick line is the “critical line,” which passes M and has the slope b / ( n c ) . For given strategy d = ( d φ , d 0 ,   d 1 ,   d 2 ) , Inequality (A34) is satisfied if and only if P( d φ , d 0 ,   d 1 ,   d 2 ) is below the critical line, showing that d = ( 1 ,   0 ,   0 ,   1 ) and ( 1 ,   0 ,   1 ,   1 ) are the only reactive strategies that are selectively favored when replacing a population of ALLD. Similarly, Inequality (A36) holds true if and only if Q( d φ , d 0 ,   d 1 ,   d 2 ) is below the critical line. Parameter values used are δ = 0.8 and b / c = 2.4 . The asterisks indicate that the corresponding element of d can be either 0 or 1.
Figure A1. Geometric representations of Inequalities (A34) and (A36) for n = 3 . Each reactive strategy d = ( d φ , d 0 ,   d 1 ,   d 2 ) is represented by two points on the x y -plane, P( d φ , d 0 ,   d 1 ,   d 2 ) and Q( d φ , d 0 ,   d 1 ,   d 2 ). The empty square represents point M, which is given by x = y = 1 1 / δ and on the line y = x ( x < 0 ) (the broken line). The thick line is the “critical line,” which passes M and has the slope b / ( n c ) . For given strategy d = ( d φ , d 0 ,   d 1 ,   d 2 ) , Inequality (A34) is satisfied if and only if P( d φ , d 0 ,   d 1 ,   d 2 ) is below the critical line, showing that d = ( 1 ,   0 ,   0 ,   1 ) and ( 1 ,   0 ,   1 ,   1 ) are the only reactive strategies that are selectively favored when replacing a population of ALLD. Similarly, Inequality (A36) holds true if and only if Q( d φ , d 0 ,   d 1 ,   d 2 ) is below the critical line. Parameter values used are δ = 0.8 and b / c = 2.4 . The asterisks indicate that the corresponding element of d can be either 0 or 1.
Games 09 00100 g0a1
The slope of the line through points Qd and M gives the right-hand side of Inequality (A36). Thus, Inequality (A36) is met if and only if Qd is below the critical line as mentioned above. From Equation (A37), it is clear that x < y holds for any Qd with d n 1 = 0 , which means that Qd with d n 1 = 0 is above the critical line, for the slope of the critical line is always smaller than one. Hence, the fixation probability of ALLD replacing a population of d is larger than 1 / N for any d having d n 1 = 0 . Figure A1 illustrates the case of n = 3 , showing that ALLD is selectively disfavored when replacing a population of d = ( 1 ,   0 ,   0 ,   1 ) , ( 1 ,   0 ,   1 ,   1 ) ,   ( 1 ,   1 ,   0 ,   1 ) , or ( 0 ,   1 ,   0 ,   1 ) . When d φ = d 0 = 0 holds true, the fixation probability of ALLD is equal to 1 / N .

References

  1. Trivers, R. The evolution of reciprocal altruism. Q. Rev. Biol. 1971, 46, 35–57. [Google Scholar] [CrossRef]
  2. Axelrod, R.; Hamilton, W.D. The evolution of cooperation. Science 1981, 211, 1390–1396. [Google Scholar] [CrossRef]
  3. Joshi, N.V. Evolution of cooperation by reciprocation within structured demes. J. Genet. 1987, 6, 69–84. [Google Scholar] [CrossRef]
  4. Boyd, R.; Richerson, P.J. The evolution of reciprocity in sizable groups. J. Theor. Biol. 1988, 132, 337–356. [Google Scholar] [CrossRef]
  5. Nowak, M.A. Stochastic strategies in the prisoner’s dilemma. Theor. Popul. Biol. 1990, 38, 93–112. [Google Scholar] [CrossRef]
  6. Nowak, M.A.; Sigmund, K. The evolution of stochastic strategies in the prisoner’s dilemma. Acta Appl. Math. 1990, 20, 247–265. [Google Scholar] [CrossRef]
  7. Press, W.H.; Dyson, F.J. Iterated prisoner’s dilemma contains strategies that dominate any evolutionary opponent. Proc. Natl. Acad. Sci. USA 2012, 109, 10409–10413. [Google Scholar] [CrossRef]
  8. Stewart, A.J.; Plotkin, J.B. Extortion and cooperation in the prisoner’s dilemma. Proc. Natl. Acad. Sci. USA 2012, 109, 10134–10135. [Google Scholar] [CrossRef]
  9. Stewart, A.J.; Plotkin, J.B. From extortion to generosity, evolution in the iterated prisoner’s dilemma. Proc. Natl. Acad. Sci. USA 2013, 110, 15348–15353. [Google Scholar] [CrossRef]
  10. Stewart, A.J.; Plotkin, J.B. Collapse of cooperation in evolving games. Proc. Natl. Acad. Sci. USA 2014, 111, 17558–17563. [Google Scholar] [CrossRef] [Green Version]
  11. Hilbe, C.; Traulsen, A.; Sigmund, K. Partners or rivals? strategies for the iterated prisoner’s dilemma. Games Econ. Behav. 2015, 92, 41–52. [Google Scholar] [CrossRef] [PubMed]
  12. Hilbe, C.; Chatterjee, K.; Nowak, M.A. Partners and rivals in direct reciprocity. Nature Hum. Behaviour. 2018, 2, 469–477. [Google Scholar] [CrossRef]
  13. Nowak, M.A.; Sigmund, K. Tit for tat in heterogeneous populations. Nature 1992, 355, 250–253. [Google Scholar] [CrossRef]
  14. Nowak, M.A.; May, R.M. Evolutionary games and spatial chaos. Nature 1992, 359, 826–829. [Google Scholar] [CrossRef]
  15. Killingback, T.; Doebeli, M. Self-organized criticality in spatial evolutionary game theory. J. Theor. Biol. 1998, 191, 335–340. [Google Scholar] [CrossRef] [PubMed]
  16. Fudenberg, D.; Harris, C. Evolutionary dynamics with aggregate shocks. J. Econ. Theory 1992, 57, 420–441. [Google Scholar] [CrossRef]
  17. Nowak, M.A.; Sasaki, A.; Taylor, C.; Fudenberg, D. Emergence of cooperation and evolutionary stability in finite populations. Nature 2004, 428, 646–650. [Google Scholar] [CrossRef] [Green Version]
  18. Kurokawa, S.; Ihara, Y. Emergence of cooperation in public goods games. Proc. R. Soc. Lond. B Biol. Sci. 2009, 276, 1379–1384. [Google Scholar] [CrossRef] [Green Version]
  19. Kurokawa, S.; Wakano, J.Y.; Ihara, Y. Generous cooperators can outperform non-generous cooperators when replacing a population of defectors. Theor. Popul. Biol. 2010, 77, 257–262. [Google Scholar] [CrossRef]
  20. Kollock, P. Social dilemmas: The anatomy of cooperation. Annu. Rev. Sociol 1998, 24, 183–214. [Google Scholar] [CrossRef]
  21. Milinski, M.; Semmann, D.; Krambeck, H.J.; Marotzke, J. Stabilizing the Earth’s climate is not a losing game: Supporting evidence from public goods experiments. Proc. Natl. Acad. Sci. USA 2006, 103, 3994–3998. [Google Scholar] [CrossRef] [PubMed]
  22. Hauert, C.; Schuster, H.G. Effects of increasing the number of players and memory size in the iterated Prisoner’s Dilemma: A numerical approach. Proc. R. Soc. Lond. B Biol. Sci. 1997, 264, 513–519. [Google Scholar] [CrossRef]
  23. Taylor, M. Anarchy and Cooperation; Wiley: New York, NY, USA, 1976. [Google Scholar]
  24. Boyd, R.; Lorberbaum, J.P. No pure strategy is evolutionarily stable in the repeated prisoner’s dilemma game. Nature 1987, 327, 58–59. [Google Scholar] [CrossRef]
  25. Boyd, R. Mistakes allow evolutionary stability in the repeated prisoner’s dilemma game. J. Theor. Biol. 1989, 136, 47–56. [Google Scholar] [CrossRef]
  26. van Veelen, M. Robustness against indirect invasions. Games Econ. Behav. 2012, 74, 382–393. [Google Scholar] [CrossRef]
  27. van Veelen, M.; García, J.; Rand, D.G.; Nowak, M.A. Direct reciprocity in structured populations. Proc. Natl. Acad. Sci. USA 2012, 109, 9929–9934. [Google Scholar] [CrossRef] [Green Version]
  28. García, J.; van Veelen, M. In and out of equilibrium I: Evolution of strategies in repeated games with discounting. J. Econ. Theor. 2016, 161, 161–189. [Google Scholar] [CrossRef]
  29. Bonner, J.T. The Social Amoeba; Princeton University Press: Princeton, NJ, USA, 2008. [Google Scholar]
  30. Yip, E.C.; Powers, K.S.; Aviles, L. Cooperative capture of large prey solves scaling challenge faced by spider societies. Proc. Natl. Acad. Sci. USA 2008, 105, 11818–11822. [Google Scholar] [CrossRef] [Green Version]
  31. Packer, C.; Scheel, D.; Pusey, A.E. Why lions form groups: Food is not enough. Am. Nat. 1990, 136, 1–19. [Google Scholar] [CrossRef]
  32. Creel, S. Cooperative hunting and group size: Assumptions and currencies. Anim. Behav. 1997, 54, 1319–1324. [Google Scholar] [CrossRef]
  33. Stander, P.E. Foraging dynamics of lions in semi-arid environment. Can. J. Zool. 1991, 70, 8–21. [Google Scholar] [CrossRef]
  34. Bednarz, J.C. Cooperative hunting Harris’ hawks Parabuteo unicinctus. Science 1988, 239, 1525–1527. [Google Scholar] [CrossRef]
  35. Rabenold, K.N. Cooperative enhancement of reproductive success in tropical wren societies. Ecology 1984, 65, 871–885. [Google Scholar] [CrossRef]
  36. Pacheco, J.M.; Santos, F.C.; Souza, M.O.; Skyrms, B. Evolutionary dynamics of collective action in N-person stag hunt dilemmas. Proc. R. Soc. Lond. B Biol. Sci. 2009, 276, 315–321. [Google Scholar] [CrossRef]
  37. Bach, L.A.; Helvik, T.; Christiansen, F.B. The evolution of n-player cooperation—Threshold games and ESS bifurcations. J. Theor. Biol. 2006, 238, 426–434. [Google Scholar] [CrossRef] [PubMed]
  38. Souza, M.O.; Pacheco, J.M.; Santos, F.C. Evolution of cooperation under n-person snowdrift games. J. Theor. Biol. 2009, 260, 581–588. [Google Scholar] [CrossRef] [PubMed]
  39. De Jaegher, K. Harsh environments and the evolution of multi-player cooperation. Theor. Popul. Biol. 2017, 113, 1–12. [Google Scholar] [CrossRef]
  40. Taylor, C.; Nowak, M.A. Transforming the dilemma. Evolution 2007, 61, 2281–2292. [Google Scholar] [CrossRef]
  41. Nowak, M.A.; Tarnita, C.E.; Wilson, E.O. The evolution of eusociality. Nature 2010, 466, 1057–1062. [Google Scholar] [CrossRef]
  42. Allen, B.; Nowak, M.A.; Wilson, E.O. Limitations of inclusive fitness. Proc. Natl. Acad. Sci. USA 2013, 110, 20135–20139. [Google Scholar] [CrossRef] [Green Version]
  43. Allen, B.; Nowak, M.A. Games among relatives revisited. J. Theor. Biol. 2015, 378, 103–116. [Google Scholar] [CrossRef] [PubMed]
  44. Kurokawa, S. Payoff non-linearity sways the effect of mistakes on the evolution of reciprocity. Math. Biosci. 2016, 279, 63–70. [Google Scholar] [CrossRef] [PubMed]
  45. Kurokawa, S. Imperfect information facilitates the evolution of reciprocity. Math. Biosci. 2016, 276, 114–120. [Google Scholar] [CrossRef] [PubMed]
  46. Peña, J.; Lehmann, L.; Nöldeke, G. Gains from switching and evolutionary stability in multi-player matrix games. J. Theor. Biol. 2014, 346, 23–33. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Archetti, M.; Scheuring, I. Review: Game theory of public goods in one-shot social dilemmas without assortment. J. Theor. Biol. 2012, 299, 9–20. [Google Scholar] [CrossRef] [PubMed]
  48. Bomze, I.; Pawlowitsch, C. One-third rules with equality: Second-order evolutionary stability conditions in finite populations. J. Theor. Biol. 2008, 254, 616–620. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Wu, B.; García, J.; Hauert, C.; Traulsen, A. Extrapolating weak selection in evolutionary games. Plos. Comput. Biol. 2013, 9, e1003381. [Google Scholar] [CrossRef] [PubMed]
  50. Slade, P.F. On risk-dominance and the ‘1/3—Rule’ in 2X2 evolutionary games. IJPAM 2017, 113, 649–664. [Google Scholar] [CrossRef]
  51. Moran, P.A.P. Random processes in genetics. Math. Proc. Camb. 1958, 54, 60–71. [Google Scholar] [CrossRef]
  52. Deng, K.; Li, Z.; Kurokawa, S.; Chu, T. Rare but severe concerted punishment that favors cooperation. Theor. Popul. Biol. 2012, 81, 284–291. [Google Scholar] [CrossRef]
  53. Zhang, C.; Chen, Z. The public goods game with a new form of shared reward. J. Stat. Mech. Theor. Exp. 2016, 10, 103201. [Google Scholar] [CrossRef]
  54. Kurokawa, S.; Ihara, Y. Evolution of social behavior in finite populations: A payoff transformation in general n-player games and its implications. Theor. Popul. Biol. 2013, 84, 1–8. [Google Scholar] [CrossRef] [PubMed]
  55. Gokhale, C.S.; Traulsen, A. Evolutionary games in the multiverse. Proc. Natl. Acad. Sci. USA 2010, 107, 5500–5504. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Hirshleifer, J. From weakest-link to best-shot: The voluntary provision of public goods. Public Choice 1983, 41, 371–386. [Google Scholar] [CrossRef]
  57. Diekmann, A. Volunteer’s dilemma. J. Confl. Resolut. 1985, 29, 605–610. [Google Scholar] [CrossRef]
  58. Fischbacher, U.; Gachter, S.; Fehr, E. Are people conditionally cooperative? Evidence from a public goods experiment. Econ. Lett. 2001, 71, 397–404. [Google Scholar] [CrossRef] [Green Version]
  59. Kocher, M.G.; Cherry, T.; Kroll, S.; Netzer, R.J.; Sutter, M. Conditional cooperation on three continents. Econ. Lett. 2008, 101, 175–178. [Google Scholar] [CrossRef] [Green Version]
  60. Herrmann, B.; Thöni, C. Measuring conditional cooperation: A replication study in Russia. Exp. Econ. 2009, 12, 87–92. [Google Scholar] [CrossRef]
  61. Martinsson, P.; Pham-Khanh, N.; Villegas-Palacio, C. Conditional cooperation and disclosure in developing countries. J. Econ. Psychol. 2013, 34, 148–155. [Google Scholar] [CrossRef] [Green Version]
  62. Takezawa, M.; Price, M.E. Revisiting “the evolution of reciprocity in sizable groups”: Continuous reciprocity in the repeated n-person prisoner’s dilemma. J. Theor. Biol. 2010, 264, 188–196. [Google Scholar] [CrossRef]
  63. Nowak, M.A.; Sigmund, K. A strategy of win-stay, lose-shift that outperforms tit-for-tat in the prisoner’s dilemma game. Nature 1993, 364, 56–58. [Google Scholar] [CrossRef] [PubMed]
  64. Kurokawa, S. Persistence extends reciprocity. Math. Biosci. 2017, 286, 94–103. [Google Scholar] [CrossRef] [PubMed]
  65. Hayden, B.Y.; Platt, M.L. Gambling for gatorade: Risk-sensitive decision making for fluid rewards in humans. Anim. Cogn. 2009, 12, 201–207. [Google Scholar] [CrossRef] [PubMed]
  66. Scheibehenne, B.; Wilke, A.; Todd, P.M. Expectations of clumpy resources influence predictions of sequential events. Evol. Hum. Behav. 2011, 32, 326–333. [Google Scholar] [CrossRef] [Green Version]
  67. Wang, Z.; Xu, B.; Zhou, H.-J. Social cycling and conditional responses in the rock-paper-scissors game. Sci. Rep. 2014, 4, 5830. [Google Scholar] [CrossRef] [PubMed]
  68. Tamura, K.; Masuda, N. Win-stay lose-shift strategy in formation changes in football. EPJ 257 Data Sci. 2015, 4, 9. [Google Scholar] [CrossRef]
  69. Kurokawa, S. Evolution of cooperation: The analysis of the case wherein a different player has a different benefit and a different cost. Lett. Evol. Behav. Sci. 2016, 7, 5–8. [Google Scholar] [CrossRef]
  70. Bowles, S.; Gintis, H. A Cooperative Secies: Human Reciprocity and its Evolution; Princeton University Press: Princeton, NJ, USA, 2011. [Google Scholar] [CrossRef]
  71. Kurokawa, S. Does imperfect information always disturb the evolution of reciprocity? Lett. Evol. Behav. Sci. 2016, 7, 14–16. [Google Scholar] [CrossRef]
  72. Kurokawa, S. Evolutionary stagnation of reciprocators. Anim. Behav. 2016, 122, 217–225. [Google Scholar] [CrossRef]
  73. Kurokawa, S. Unified and simple understanding for the evolution of conditional cooperators. Math. Biosci. 2016, 282, 16–20. [Google Scholar] [CrossRef]
  74. Kurokawa, S. The extended reciprocity: Strong belief outperforms persistence. J. Theor. Biol. 2017, 421, 16–27. [Google Scholar] [CrossRef] [PubMed]
  75. Kurokawa, S.; Ihara, Y. Evolution of group-wise cooperation: Is direct reciprocity insufficient? J. Theor. Biol. 2017, 415, 20–31. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Functional forms of f ( k ) for various payoff assumptions. The horizontal and vertical axes represent k , the number of cooperating individuals in a given round among the n group members, and f ( k ) , a function of k as defined by Equation (13), respectively. Parameter values used are N = 400 and n = 30 , (a) The conventional linear payoff assumption, Equations (19) and (20); b = 24 and c = 1 , (b) A non-linear payoff assumption, Equations (42) and (43); (c) A non-linear payoff assumption, Equations (20) and (52). b = 1 .
Figure 1. Functional forms of f ( k ) for various payoff assumptions. The horizontal and vertical axes represent k , the number of cooperating individuals in a given round among the n group members, and f ( k ) , a function of k as defined by Equation (13), respectively. Parameter values used are N = 400 and n = 30 , (a) The conventional linear payoff assumption, Equations (19) and (20); b = 24 and c = 1 , (b) A non-linear payoff assumption, Equations (42) and (43); (c) A non-linear payoff assumption, Equations (20) and (52). b = 1 .
Games 09 00100 g001
Figure 2. The effects of w and δ on the identity of the best reactive strategy. The horizontal axis represents w (0.001 to 0.161 with the interval of 0.02) and the vertical axis is δ (0.05 to 0.95 with the interval of 0.05. The fixation probabilities of strategies are obtained without assuming weak selection (i.e., using Equation (8). The parameter values used are N = 50 , n = 10 , b = 9 , and c = 1 .
Figure 2. The effects of w and δ on the identity of the best reactive strategy. The horizontal axis represents w (0.001 to 0.161 with the interval of 0.02) and the vertical axis is δ (0.05 to 0.95 with the interval of 0.05. The fixation probabilities of strategies are obtained without assuming weak selection (i.e., using Equation (8). The parameter values used are N = 50 , n = 10 , b = 9 , and c = 1 .
Games 09 00100 g002
Figure 3. Relative fixation probabilities of all reactive strategies for the linear public goods game with n = 4 . The horizontal axis represents ( N ρ d ,   ALLD 1 ) / w , which is positive when ρ d ,   ALLD > 1 / N (i.e., the fixation probability of a single mutant of strategy d , when introduced in a population of ALLD, exceeds what is expected under neutral evolution), and is negative when the inequality is reversed. The strategies are aligned in the order of the fixation probabilities, which are calculated using Equation (9). Results for different value of δ are shown: (a) δ = 0.8 , (b) δ = 0.6 , and (c) δ = 0.4 . Other parameter values are N = 100 , b = 3.5 , and c = 1 . The asterisks indicate that the corresponding element of d can be either 0 or 1. The order of the strategies is affected by δ .
Figure 3. Relative fixation probabilities of all reactive strategies for the linear public goods game with n = 4 . The horizontal axis represents ( N ρ d ,   ALLD 1 ) / w , which is positive when ρ d ,   ALLD > 1 / N (i.e., the fixation probability of a single mutant of strategy d , when introduced in a population of ALLD, exceeds what is expected under neutral evolution), and is negative when the inequality is reversed. The strategies are aligned in the order of the fixation probabilities, which are calculated using Equation (9). Results for different value of δ are shown: (a) δ = 0.8 , (b) δ = 0.6 , and (c) δ = 0.4 . Other parameter values are N = 100 , b = 3.5 , and c = 1 . The asterisks indicate that the corresponding element of d can be either 0 or 1. The order of the strategies is affected by δ .
Games 09 00100 g003aGames 09 00100 g003b
Table 1. The payoff matrix of the general n-player game.
Table 1. The payoff matrix of the general n-player game.
Number   of   a   Individuals   among   the   n 1   Opponents
Strategy of the focal individual n 1 n 2 n 3 1 0
A a 1 a 2 a 3 a n 1 a n
B b 1 b 2 b 3 b n 1 b n
Table 2. A list of the ten best reactive strategies for n = 10 under the linear payoff assumption of Equations (19) and (20). Vector d and the corresponding value of ( N ρ d , ALLD 1 ) / w are given. Parameter values are N = 100 , b = 8 , c = 1 , and δ = 0.8 .
Table 2. A list of the ten best reactive strategies for n = 10 under the linear payoff assumption of Equations (19) and (20). Vector d and the corresponding value of ( N ρ d , ALLD 1 ) / w are given. Parameter values are N = 100 , b = 8 , c = 1 , and δ = 0.8 .
Ranking d ( N ρ d ,   A L L D 1 ) / w
1(1,0,0,0,0,0,0,1,1,1,1)40.39
2(1,0,0,0,0,0,1,1,1,1,1)38.14
3(1,0,0,0,0,0,0,0,1,1,1)36.35
4(1,0,0,0,0,0,1,0,1,1,1)34.1
5(1,0,0,0,0,1,0,1,1,1,1)31.85
6(1,0,0,0,0,0,0,1,0,1,1)30.06
7(1,0,0,0,0,1,1,1,1,1,1)29.59
8(1,0,0,0,0,1,0,0,1,1,1)27.81
8(1,0,0,0,0,0,1,1,0,1,1)27.81
10(1,0,0,0,0,0,0,0,0,1,1)26.03

Share and Cite

MDPI and ACS Style

Kurokawa, S.; Wakano, J.Y.; Ihara, Y. Evolution of Groupwise Cooperation: Generosity, Paradoxical Behavior, and Non-Linear Payoff Functions. Games 2018, 9, 100. https://doi.org/10.3390/g9040100

AMA Style

Kurokawa S, Wakano JY, Ihara Y. Evolution of Groupwise Cooperation: Generosity, Paradoxical Behavior, and Non-Linear Payoff Functions. Games. 2018; 9(4):100. https://doi.org/10.3390/g9040100

Chicago/Turabian Style

Kurokawa, Shun, Joe Yuichiro Wakano, and Yasuo Ihara. 2018. "Evolution of Groupwise Cooperation: Generosity, Paradoxical Behavior, and Non-Linear Payoff Functions" Games 9, no. 4: 100. https://doi.org/10.3390/g9040100

APA Style

Kurokawa, S., Wakano, J. Y., & Ihara, Y. (2018). Evolution of Groupwise Cooperation: Generosity, Paradoxical Behavior, and Non-Linear Payoff Functions. Games, 9(4), 100. https://doi.org/10.3390/g9040100

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop