Next Article in Journal
The Various Definitions of Multiple Differentiability of a Function f: ℝn→ ℝ
Next Article in Special Issue
A Differential Game with Random Time Horizon and Discontinuous Distribution
Previous Article in Journal
A New Algorithm for the Common Solutions of a Generalized Variational Inequality System and a Nonlinear Operator Equation in Banach Spaces
Previous Article in Special Issue
Secretary Problem with Possible Errors in Observation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prioritised Learning in Snowdrift-Type Games

by
Maria Kleshnina
1,*,
Sabrina S. Streipert
2,
Jerzy A. Filar
3 and
Krishnendu Chatterjee
1
1
Institute of Science and Technology Austria (IST Austria), Am Campus 1, 3400 Klosterneuburg, Austria
2
Department of Mathematics and Statistics, McMaster University, 1280 Main St W, Hamilton, ON L8S 4L8, Canada
3
Centre for Applications in Natural Resource Mathematics, School of Mathematics and Physics, University of Queensland, St Lucia, QLD 4072, Australia
*
Author to whom correspondence should be addressed.
Mathematics 2020, 8(11), 1945; https://doi.org/10.3390/math8111945
Submission received: 7 October 2020 / Revised: 24 October 2020 / Accepted: 26 October 2020 / Published: 4 November 2020
(This article belongs to the Special Issue Statistical and Probabilistic Methods in the Game Theory)

Abstract

:
Cooperation is a ubiquitous and beneficial behavioural trait despite being prone to exploitation by free-riders. Hence, cooperative populations are prone to invasions by selfish individuals. However, a population consisting of only free-riders typically does not survive. Thus, cooperators and free-riders often coexist in some proportion. An evolutionary version of a Snowdrift Game proved its efficiency in analysing this phenomenon. However, what if the system has already reached its stable state but was perturbed due to a change in environmental conditions? Then, individuals may have to re-learn their effective strategies. To address this, we consider behavioural mistakes in strategic choice execution, which we refer to as incompetence. Parametrising the propensity to make such mistakes allows for a mathematical description of learning. We compare strategies based on their relative strategic advantage relying on both fitness and learning factors. When strategies are learned at distinct rates, allowing learning according to a prescribed order is optimal. Interestingly, the strategy with the lowest strategic advantage should be learnt first if we are to optimise fitness over the learning path. Then, the differences between strategies are balanced out in order to minimise the effect of behavioural uncertainty.

1. Introduction

In many situations, ability to learn becomes important for survival: how fast and effectively individuals explore their surroundings and learn how to react to them determines their survival chances. Some environments and situations might require specific skills or rapid adaptations whereas others might just require general awareness. Learning and adaptation are remarkable features of life in natural environments that are prone to frequent changes. The process of learning can be considered as a continuous multidimensional process where several skills can be learned in parallel. However, reducing the time step to a very small interval, we focus on learning only one skill at a time. In this study, we introduce a model of multidimensional stepwise learning by assuming that the learning process can be subdivided into smaller time steps during which we only learn one skill.
Evolutionary games were proposed to answer questions of the effects of natural selection on populations. Here, a game happens at a higher level: the game is not played directly by individuals but rather by strategies or skills competing for survival [1,2]. In such settings, learning the best and more effective strategy first becomes critical for individuals, especially under resource limitation [3]. Learning in the evolutionary settings often assumes that learning evolves in parallel with the evolution itself [4]. In this paper, we consider the case when the system is perturbed after reaching its stable equilibrium. Finding the most efficient learning path to return to the previous equilibrium is a challenge. In this paper, we suggest to look at this problem from the evolutionary perspective where an optimal learning path is dictated by the chance of survival. This is achieved through maximisation of fitness depending on the learning path an individual takes.
Behavioural stochasticity is a popular object of study for game-theorists. First, the concept of “trembling hands” [5] was suggested as an approach to model players’ mistakes during the strategies’ execution with some small probability. In addition, mutations can be interpreted as reactions to environmental changes (see for instance [6,7,8] for studies on mutations). In artificial intelligence and economics this idea is of a particular interest since individuals can be imperfect and incapable of executing their strategies (examples of such studies are [9,10,11]). However, we do not restrict our analysis to mistakes that are random noise, but, instead, allow for biased mistakes and assume that skills can be learnt at different rates. For this, we utilise the notion of incompetence that means that individuals make mistakes with certain probabilities, which was first introduced in [12,13] and extended in [14]. Hence, learning is defined as a process of improving competence and reducing the probabilities of mistakes. Here, we propose the notion of prioritised learning when the priority in the order is determined by the skills’ relative advantages. That is, we are interested in the mechanism of discovering an optimal order according to which skills should be learned. This mechanism relies on the interplay between relative fitness and learning advantages and the notion of prioritised learning which aims to balance out these differences between the strategies.
We focus on the question in which order species should learn their skills to have an evolutionary advantage. The process of re-learning the effective strategy can be considered as a continuous multidimensional problem. However, reducing the time step to a very small interval, we focus on only one skill. We consider the case of the coexistence of two strategies: to cooperate and to defect. Mathematically, this can be described by the Snowdrift (also known as Chicken or Hawk-Dove) game where two strategies interact and both appear in the behavioural traits of individuals [15].
The question of how cooperation evolves in communities is a rich field of study for mathematicians and biologists. Enforcement, punishment and reciprocity are the mechanisms that can sustain stable cooperation [16,17,18,19]. The Prisoners’ Dilemma is the well-known example of the problem of cooperation [20,21]. It is a strict game where sustaining cooperation becomes a challenge. However, it can be relaxed if we assume that benefit gained from cooperation can be received by both players and the costs shared (see [22] for the review). The resulting game has a form of the Snowdrift game, where a stable mixed equilibrium exists [23]. We point out that the focus of our study is not the evolution of cooperative mechanisms as such. Learning to cooperate is a fruitful field proposing many interesting results [24,25,26]. However, here we are more interested in understanding of how the system should optimally evolve back to an equilibrium state once it was disturbed. Specifically, we shall focus on a Snowdrift type game in order to determine which of the strategies (cooperation or defection) should be learnt first.
When speaking of the optimality of a learning path, we expect that the benefit from this strategy has an impact on the overall population’s fitness [27]. We note that learning and replicator timescales are decoupled. In fact, we require that the learning timescale is much slower than reproduction. This can be seen in view of a behavioural adaptation that might take longer time to happen. In classic settings of the replicator dynamics, the dynamics’ time scale is equal to the reproduction time scale. That is, every time step is the length of one generation. However, when modelling behaviour of individuals, this reproduction time scale represents an interaction time scale. Therefore, the behavioural change might take several interactions to be achieved. Hence, naturally, we assume that individuals are learning slower than they interact. For example, in the incompetent version of the Snowdrift game, individuals go through multiple interactions which may change their behaviour from cooperating to defecting.
We show that measuring the fitness over the learning path depends on the order of learning along the path and the extent to which strategies are learnt. Our results demonstrate that it might be preferable to learn a skill prone to higher probabilities of mistakes first and leverage its learning advantage. This suggests that in the environmental or other changes those strategies that are most disruptive must be adapted to more quickly if they are to survive at all. If two skills are equivalent in their relative strategic advantages, we show that both skills are also equivalent with respect to the order of learning. Counter-intuitively, we show that these relative advantages can still be identical even if the relative fitness advantages of the skills are significantly different suggesting that the evolution is trying to balance out mistakes. We conjecture that not only the fitness of the skill has to be taken into account but also the degree of incompetence.
In the following section we set up the model and define the notions of relative fitness, learning and strategic advantages. Then, in Section 3 we define a fitness-over-learning objective function measuring how fitness of the population improves over the learning path taken. After that, we proceed to two cases: (a) the case when skills are identical in their strategic advantage and no prioritised learning is needed in Section 4, and (b) the case when skills are distinct in their advantages in Section 5. These two cases demonstrate the notion of hierarchy in the learning order and how mistakes affect the evolution.

2. Learning Setup

Typically, a population of species acquires a set of skills that they are required to learn either while being young or while adapting to new environmental conditions. Let us suppose that this set consists of only two essential skills both of which are needed for survival (or for stable coexistence in the population). Hence, we consider two specific strategies that need to be learnt: cooperation and defection. In the evolutionary context, defection can be interpreted as aggression to capture (rather than share) a resource such as food or territory. We utilise the notion of replicator dynamics in order to describe the evolution of interactions among individuals in the population [28,29]. Let p : = p ( t ) [ 0 , 1 ] be the frequency of cooperators at (evolutionary) time t. Then, the dynamics can be expressed as
p ˙ = ( f C ϕ ) p ,
where f C is fitness of cooperators and ϕ is the mean fitness function defined as ϕ : = f C p + f D ( 1 p ) , since, ( 1 p ) is the fraction of defectors. For the purpose of this paper, we use a linear form of the fitness functions, as in [14], which simplifies to
p ˙ = ( f C ϕ ) p = ( f C f D ) p ( 1 p ) .
We shall note that if both cooperation and defection strategies would be required to some extent, then both strategies might be required to coexist at equilibrium. Such an equilibrium usually characterises the Snowdrift game, which is also referred to as a anti-coordination game. Since both strategies are the best response to the opposite strategy, at equilibrium, both strategies coexist, hence, securing some stable level of cooperation. We shall construct a reward matrix R for such a game as
Cooperate   Defect   Cooperate   Defect   B C 2 B C B 0
where B is the benefit and C is the cost of cooperation. In order to simplify our analysis, we apply the linear transformation from [30] and subtract the diagonal elements from the corresponding columns, as it does not affect dynamics’ behaviour. Then, we can consider a reduced form of the matrix given by
Cooperate   Defect   Cooperate   Defect               0                 a         b                 0             ,
where a = B C and b = C 2 . As we want to set our game to be a Snowdrift game, we assume that B > C > 0 , because then both strategies will stably coexist at the equilibrium p ^ = ( a ^ , b ^ ) and their frequencies will be given by
a ^ : = a a + b = 2 B 2 C 2 B + C and b ^ : = b a + b = C 2 B + C .
Then, in the context of learning a set of skills, both skills will be required to be learned. The question we address is: which one should be learnt first?
In order to model learning these skills in the game, we utilise the notion of incompetence. This concept was first introduced for classic games in 2012 [12] and extended to evolutionary games in 2018 [14]. Here, individuals choose a strategy to play but have non-zero probabilities of executing a different strategy. Such mistakes are a manifestation of the incompetence of players. Mathematically, it is described by the matrix of incompetence, Q, that evolves as individuals are learning. This matrix consists of conditional probabilities q i j determining the probability of executing strategy j given that strategy i was chosen. The matrix has the form
Q = q 11 q 12 q 21 q 22 .
A schematic representation of the interaction under incompetence can be found in Figure 1A. As a measure of learning we use parameters x and y for each strategy. Thus, each strategy can be learned at a different pace, which sets this work apart from the existing literature (see [12,31]). Then, the matrix of incompetent parameters has the form
X = x 0 0 y ,
where x , y [ 0 , 1 ] . We assume that strategies can be learnt and, hence, the incompetence can decrease from some initial level of propensity to make mistakes. Thus, the learning process is described by the equation
Q ( X ) = ( I X ) S + X ,
where S represents the starting level of incompetence. If x = y = 0 , then X = 0 and
Q ( 0 ) = S = α 1 α 1 β β .
Full competence corresponds to the case x = y = 1 , where Q = I , the identity matrix. Each strategy has its own measure of incompetence level and its own time needed for it to be mastered. Then, the new incompetent game reward matrix is defined as
R ( X ) = R ( x , y ) = Q ( X ) × R × Q ( X ) T .
Given the new reward matrix, we require some basic assumptions on parameters in (3)–(5) to be satisfied in order for the game to avoid bifurcations in the replicator dynamics. Specifically:
  • a , b > 0 : this ensures that it is beneficial to learn both strategies.
  • α > a ^ , β > b ^ : this is necessary for the new incompetent game to have a stable mixed equilibrium point for any level of incompetence, x , y [ 0 , 1 ] [14].
  • α + β > 1 : this condition also guarantees existence of the stable mixed equilibrium [14].
If parameters of the game do not meet these conditions, then there will be some values of the incompetence parameters for which the system (1) undergoes bifurcation. This would lead to situations where one of the skills is not beneficial to be learnt due to the fact that it is dominated. Hence, learning the beneficial skill would be an obvious answer to the question of which skill to learn first. However, we are interested in the case when the optimal learning path involves both skills.

3. Maximising Fitness over Learning

The optimality of a learning path can be determined in many ways. In the replicator dynamics with symmetric payoff matrices, interactions follow the evolutionary path that maximises the overall population’s fitness [32]. We focus on the fitness function at the equilibrium state, which implies that the population attains a steady-state faster than incompetence parameters change. Technically speaking, we assume either very long timescales of learning or else very fast convergence to steady state, or both. Hence, it is sufficient to consider the mean fitness function [29] of the Snowdrift game which has the following form
ϕ = p ^ R p ^ T = a b a + b .
Then, in our new incompetent game the mean fitness ϕ ( X ) = p ^ R ( X ) p ^ T can be shown to be
ϕ ( X ) = ( a + b ) ( α a ^ ) ( β b ^ ) + y ( 1 β ) α a ^ + x ( 1 α ) ( β b ^ ) + x y ( 1 α ) ( 1 β ) .
We shall note that in fact any nontrivial learning path will improve the mean fitness at the equilibrium. This follows from the fact that for any vector x = ( x , y ) with entries in ( 0 , 1 ) the following holds
x H ϕ x T = 2 ( a + b ) x y ( 1 α ) ( 1 β ) 0 ,
where H ϕ is a Hessian of ϕ ( X ) . Since we would like to find an optimal learning path that maximises the fitness, it is convenient to consider a new re-scaled fitness function. For that, we choose
ϕ ˜ ( X ) : = ϕ ( X ) ( a + b ) ( 1 α ) ( 1 β ) = a ˜ b ˜ + b ˜ x + a ˜ y + x y ,
where new parameters a ˜ and b ˜ are
a ˜ : = α a ^ 1 α , b ˜ : = β b ^ 1 β .
An important note for further analysis is the understanding of fitness and learning advantages. Given that the relative fitness of each strategy is positive, that is, a , b > 0 , we say that the strategy with a higher relative fitness obtains a fitness advantage. In addition, we define a learning advantage of a strategy in a very special way. This advantage arises when the strategy is accompanied by a higher probability of making a mistake in its execution. In such a case reducing the corresponding level of incompetence offers a greater opportunity for improvement, thereby constituting a learning advantage.
Next, note that the parameters a ˜ and b ˜ introduced in (8) are closely connected with the above definitions of fitness and learning advantages. Indeed, their difference allows us to capture the relative tradeoffs between these two types of advantages. Hence, we shall say that the new parameter δ : = a ˜ b ˜ measures the relative strategic advantage of cooperators against defectors. That is, if δ > 0 , cooperators have an advantage. We illustrate these concepts in Table 1. Next, we shall demonstrate how this notion affects the optimal learning path with respect to the population’s mean fitness.
However, we need to take into account possible limitations of the learning pace. We assume that only one skill can be learnt to some degree and two skills cannot be learnt simultaneously. Thus, learning is achieved in a discrete manner (see Figure 1B). The question is what is the optimal learning order and what are the switching points. This will be determined by defining the learning path as a curve C on the learning space L such that it starts at ( 0 , 0 ) and ends at ( 1 , 1 ) .
Definition 1.
Define the learning space of an incompetence game L as the domain of the incompetence matrices Q ( X ) from (5) given by the set of all 2 × 2 stochastic matrices.
Definition 2.
Define a learning path for an incompetence game as a curve C : I 2 L on the learning space L such that C ( 0 ) = 0 = ( 0 , 0 ) T and C ( 1 ) = 1 = ( 1 , 1 ) T .
A learning path C can be a smooth curve or a stepwise path, depending on whether learning is continuous or discrete. We shall consider only stepwise learning paths. This is a natural restriction because: at a small time-scale only one skill can be learned at any given time. However, we will also study the case when the number n , which approaches a smooth learning curve.
Definition 3.
A stepwise learning path of order n, C n = { ( x k , y k ) : 1 k n } , is a stepwise curve in the learning space L , connecting the n points ( x k , y k ) k = 1 n , that satisfies
(a)
x k + 1 > x k , 1 k n 1 ,
(b)
y k + 1 > y k , 1 k n 1 ,
(c)
x 1 = y 1 = 0 ,
(d)
x n = y n = 1 .
Conditions ( a ) and ( b ) imply Δ x k , Δ y k > 0 for 1 k n 1 . The path segment from ( x k , y k ) to ( x k + 1 , y k + 1 ) could consist of the following sequence ( x k , y k ) , ( x k + 1 , y k ) , ( x k + 1 , y k + 1 ) , in which case we say that the x direction was taken. The alternative path segment is ( x k , y k ) , ( x k , y k + 1 ) , ( x k + 1 , y k + 1 ) , indicating that the y direction was chosen first.
Definition 4.
The set of all (alternating) stepwise curves of order n that satisfy conditions (a)–(d) is called the learning set of order n and is denoted by L n .
Here, we focus on alternating stepwise paths where the first direction determines the remaining path. Consequently, an n-step learning path is described by the points { ( x k , y k ) } k = 1 n satisfying (a)–(d) above, resulting in two possible learning paths corresponding to the direction of the first step. Let Φ x ( C n ) be the fitness-over-learning in the x direction for the learning path C n L n given by
Φ x ( C n ) = k = 1 n 1 x k x k + 1 ϕ ( s , y k ) d s + y k y k + 1 ϕ ( x k + 1 , s ) d s .
Let Φ y ( C n ) be the fitness-over-learning in the y direction for the learning path C n L n defined as
Φ y ( C n ) = k = 1 n 1 y k y k + 1 ϕ ( x k , s ) d s + x k x k + 1 ϕ ( s , y k + 1 ) d s .
That is, for a given n N , we have two objective functions: Φ x and Φ y . Finding their maxima separately yields an optimal learning path with the optimal direction. In what follows, we define the optimal learning paths in the x and y directions and the overall optimal learning path that maximises the population’s measure of fitness over learning.
Definition 5.
The optimal learning path C n * L n with respect to the population’s mean fitness function ϕ ( X ) : L n R is such that it satisfies the equation
C n * = C n * x , i f Φ x ( C n * x ) > Φ y ( C n * y ) C n * y , i f Φ x ( C n * x ) < Φ y ( C n * y )
where C n * x , C n * y are the optimal paths in the x and y direction, respectively. When Φ x * = Φ y * , both C n * x and C n * y are optimal paths, where Φ j * = Φ j * ( C n * j ) , j { x , y } .
The superscript (respectively, subscript) indicates the direction of the path. That is, if the x direction is optimal, that is Φ x * > Φ y * , then C n * = C n * x . Otherwise, if Φ x * < Φ y * , then C n * = C n * y . When Φ x * = Φ y * , both directions x and y are optimal and it does not matter which one we take. Hence, we can define prioritised learning in these settings.
Definition 6.
We say that there exists prioritised learning for Φ among stepwise learning paths of order n, if there exists C n * such that one of the directions is preferable over the other. That is, Φ x ( C n ) * Φ y ( C n ) * .
Given the structure of the fitness Functions (7) and (9)–(10) we can explicitly derive Φ x and Φ y for the learning paths in the x and y directions. However, we will show that the optimal direction of learning can be determined simply by the sign of the relative strategic advantage, δ .

4. No Strategic Advantages

Throughout this section, we assume that no strategy has a relative strategic advantage ( δ = 0 ). We shall first note that in this case both objective functions, Φ x and Φ y , exhibit a symmetry relation.
Theorem 1.
If δ = 0 , then there is no difference in the direction of optimal learning, that is, Φ x * = Φ y * .
Hence, if there is no relative strategic advantage in the game ( δ = 0 ), the order of learning does not affect the fitness of the population. It is therefore sufficient to calculate only one path that maximises the fitness-over-learning of the population. In the following proposition we show that this learning path has a remarkably simple form.
Proposition 1.
If δ = 0 , then the unique optimal stepwise learning path of order n in the x direction, C n * x = { ( x k * , x , y k * , x ) : 1 k n } L n , is given by
x k * , x = 0 , k = 1 2 k 3 2 n 3 , 2 k n , y k * , x = 2 k 2 2 n 3 , 1 k n 1 1 , k = n .
See Mathematical Appendix A for the proofs of Theorem 1 and Proposition 1. Interestingly, the optimal solution for the x direction yields x k * , x , y k * , x such that y k * , x x k * , x , k . That is, each step in the y direction, y k is greater than the step in the x direction, x k .
To analyse how increasing the number of steps is changing the objective function, we consider the rate of change of the fitness over learning function at the optimal solution, that is
Δ Φ x * = Φ x * ( n + 1 ) Φ x * ( n ) .
After substituting the optimal solution (11) into (12) we obtain
Δ Φ x * = 4 3 ( 2 n 3 ) > 0 , n > 2 .
Due to symmetry, when δ = 0 , it follows that Δ Φ x * = Δ Φ y * .
Therefore, the smaller the learning steps, the greater the benefit. However, the marginal increases tend to 0 as n becomes large. Arguably, this illustrates the “law of diminishing returns” of stepwise learning.
Let us consider two following games, which will be called Examples 1 and 2:
R 1 = 0 2 2 0 , S 1 = 0.5 0.5 0.5 0.5 and R 2 = 0 9 1 0 , S 2 = 0.9 0.1 0.9 0.1 .
In terms of fitnesses, these examples are different: the fixed points of the replicator equations are ( 0.5 , 0.5 ) and ( 0.9 , 0.1 ) , respectively. Hence, fitness of strategy 1 is higher in the second game. However, setting α 1 = β 1 = 0.5 and α 2 = 0.9 , β 2 = 0.1 , we have relative strategic advantages in examples 1 and 2 both equal to 0 ( δ 1 = δ 2 = 0 ). Then, S 1 and S 2 equalise the strategies. The high probability of mistakes for strategy 2 in R 2 signals that it is more disrupted by incompetence. This makes the optimal learning paths for these two games identical (see Figure 2A).
Next we shall consider the case when δ 0 . In this case, one of the strategies has a relative strategic advantage, depending on the sign of δ . We show that the order of the learning path is now important and influences the value of the fitness-over-learning function, which we call prioritised learning.

5. Prioritised Learning

First, we recall the notion of prioritised learning used henceforth. By Definition 6, there exists prioritised learning for Φ among stepwise learning paths of order n, if there exists C n * such that one of the directions is preferable over the other, along that path.That is, Φ x ( C n ) * Φ y ( C n ) * . We shall next characterise an optimal solution in the x direction.
Proposition 2.
Let n N , n 2 . If n is such that δ < 1 2 ( n 2 ) , then the x direction is preferred and the optimal learning path of order n is given by
x k * , x = 0 , k = 1 δ + ( 1 + δ ) 2 k 3 2 n 3 , 2 k n , y k * , x = ( 1 + δ ) 2 k 2 2 n 3 , 1 k n 1 1 , k = n .
We refer the reader to Mathematical Appendix A for more details. Next, we provide conditions for which the y direction defines the optimal learning path.
Proposition 3.
Let n N , n 2 . If n is such that δ > 1 2 ( n 2 ) , then the y direction is preferred and the optimal learning path of order n is given by
y k * , y = 0 , k = 1 δ + ( 1 δ ) 2 k 3 2 n 3 , 2 k n , x k * , y = ( 1 δ ) 2 k 2 2 n 3 , 1 k n 1 1 , k = n .
Hence, depending on the value of δ , the optimal learning path has different directions. The threshold for δ is equal to 1 2 ( n 2 ) , which for sufficiently large number of steps is nearly 0. However, in the proof of Theorem 2 (see Mathematical Appendix A), we show that it is the sign of δ that determines the direction of learning.
Theorem 2.
The direction of the optimal learning path is determined by the sign of δ: for δ > 0 the y direction is optimal and for δ < 0 the x direction is optimal.
For δ = 0 the optimal learning path represents n 2 equally distributed steps and the direction of the first step does not affect the fitness-over-learning. Note that if δ 0 , the first and last steps for each direction are
y 2 * , y = δ + 1 δ 2 n 3 , x n 1 * , y = 1 y 2 * , y ; x 2 * , x = δ + 1 + δ 2 n 3 , y n 1 * , x = 1 x 2 * , x .
Thus, the first step of the learning path aims to adjust fitness and learning advantages between the two strategies. However, in the interior of the learning space, the path still takes n 2 equally distributed steps. In Figure 3 we displayed the objective function Φ for different positive values of δ and different number of steps n. The images below the colormap enhance the changes in Φ * as the number of steps n varies. The difference in the values of Φ * with respect to n is marginal, hence, below we zoom into several values of δ . For δ < 1 , Φ y * increases in n, suggesting that taking more steps is beneficial. However, Φ y * decreases for δ > 1 , suggesting one-step learning of the skill. We show this in the next result that follows immediately from Propositions 2, 3, and Theorem 2.
Corollary 1.
For | δ | > 1 , we obtain two cases:
(i) 
If δ 1 , then the optimal learning curve is a one-step function in the y direction.
(ii) 
If δ 1 , then the optimal learning curve is a one-step function in the x direction.
If | δ | > 1 , then the optimal solution from (13) and (14) is only feasible for n = 2 . However, if δ is positive and less than 1, the greatest fitness-over-learning is achieved for a smooth learning path along the line y = δ + x . This can be seen as a consequence of allowing n to approach infinity.
Corollary 2.
For 0 < δ < 1 , the optimal step-wise learning path { ( x k * , y , y k * , y ) } i = 1 n in the y direction as n , follows the relation
y * , y = δ , x * , y = 0 δ + x * , y , x * , y ( 0 , 1 δ ] 1 , x * , y ( 1 δ , 1 ] .
The same relation between x * and y * can be obtained for the optimal solution in the x direction when δ < 0 and n . In that case, when we initiate learning in the x direction, we start with x * , x = δ for y * , x = 0 and continue to follow the relationship y * , x = δ + x * , x for x * , x ( 0 , 1 + δ ) . We demonstrate the effect of the relative strategic advantage on the optimal learning path, by considering the game with the fitness matrix R 3 and the incompetence matrix S 3 , referred to as Example 3:
R 3 = 0 a 1 0 , S 3 = 0.9 0.1 0.9 0.1 .
Strategy 2 obtains a learning advantage ( β 3 = 0.1 α 3 = 0.9 ). We give strategy 1 a fitness advantage by selecting three values for a: 5, 7 and 9, which result in δ 3 0.74 , 0.28 and 0, respectively. Then, smaller values of a result in the larger first step (see Figure 2B).
The next natural question to ask would be what is the influence of the number of steps we make. Computing the exact form of Φ x * and Φ y * at the optimal solutions and taking their rate of change in n yields that
Δ Φ x * = 2 ( δ + 1 ) 3 3 ( 2 n 3 ) and Δ Φ y * = 2 ( δ + 1 ) 3 3 ( 2 n 3 ) ,
which are positive for any n 2 . For δ 0 we can observe negative Φ x * identifying that the x direction is no longer preferred. This indicates that the bigger the difference between skills, the lower potential fitness-over-learning that can be gained. In this sense, the skill with higher incompetence might reduce the fitness as it requires some investments for the skills to be learnt.
Our learning scheme allows for an adjustment of individuals’ behaviour in case of any disruption leading to behavioural mistakes. The number of steps required for such an adjustment can be as low as 2 allowing for a quick reaction to system’s uncertainty. Moreover, such an adjustment does not require for the system to stop interactions for learning. Individuals can continue interacting with their group-mates while their behavioural mistakes are reduced and fitness is maximised.

6. Conclusions

In this paper, we considered the evolutionary game where two skills coexist in a mixed equilibrium, and hence are both required. This is a key assumption as we aimed to answer the question: If both strategies are important, then how do we learn them in an optimal way? We introduced a fitness-over-learning function which measures the improvement in fitness of the population over the learning path that was taken. This function relies on both performance of the strategy and its rate of mistakes.
The naive suggestion would be that the most advantageous skill in terms of fitness has to be learnt first. However, the strategy with lower relative strategic advantage is learnt first in the optimal learning path. We conjecture that this adjusts the difference between the skills and then, once they are comparable, optimal learning suggests to learn both skills with equal rates. These findings indicate that once disrupted, selection tries to recover the most affected strategies first even if their fitness is not the highest. Nonetheless, if the fitness difference is high enough to overcome the effect of incompetence, then the optimal learning will demand that the better strategy is learned first. Another possible interpretation would be to consider the mixed equilibrium as mixed strategies used by players. Then, by learning the less-advantageous strategy, individuals are reaching the nearest optimal mixed strategy.
Importantly, we parametrised the notion of strategic advantage of cooperation versus defection with a single quantity denoted by δ , which captures the tradeoffs between fitness and propensity to make execution errors in these two modes of behaviour. Interestingly, we showed that this quantity has a critical threshold absolute value of 1. Namely, if | δ | < 1 then our Corollary 1 implies that learning by many small steps is preferable to learning by fewer large steps. Arguably, this captures the belief that complex skills are best learned incrementally. However, if | δ | > 1 , then Corollary 1 shows that coexistence is preserved by one of only two possible learning paths: (a) full learning first in the x direction, followed by full learning in the y direction; or (b) the other way around. This suggests that a sufficiently large strategic advantage of cooperation over defection (or the converse) eliminates the luxury of incremental learning.
In addition, the number of steps in the learning path maximising the fitness is not bounded. Indeed, taking many small learning steps improves the fitness we observe. However, as demonstrated in Figure 3, there may exist a number of steps n * after which the increase in the objective function seems insignificant. Hence, we can determine a sufficient number of steps to achieve a target level of the fitness-over-learning function in applications.
Overall, the learning scheme proposed in this paper can be used to correct behavioural uncertainty when the system already reached its equilibrium but was disrupted. However, our formulation has its limitations. The main limitation is that we allow for only one skill to be learnt at a time. However, restrictiveness of this assumption decreases with increasing number of steps. The second limitation of our scheme is that the direction of the learning path can only be chosen at the very beginning and cannot be changed while individuals are learning. While it may be more natural to permit the learning direction to change at each step, it would also require more resources to be spent on learning. Therefore, the cost of learning would need to be taken into account. Such extensions can be studied in future research.

Author Contributions

Conceptualisation, M.K., S.S.S. and J.A.F.; validation, M.K., S.S.S., J.A.F. and K.C.; formal analysis, M.K. and S.S.S.; investigation, M.K. and S.S.S.; resources, K.C.; writing—original draft preparation, M.K.; writing—review and editing, S.S.S., J.A.F. and K.C.; visualisation, M.K. and S.S.S.; supervision, J.A.F. and K.C.; project administration, M.K.; funding acquisition, M.K., J.A.F. and K.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the European Union’s Horizon 2020 research and innovation program under the Marie Sklodowska-Curie Grant Agreement #754411, the Australian Research Council Discovery Grants DP160101236 and DP150100618, and the European Research Council Consolidator Grant 863818 (FoRM-SMArt).

Acknowledgments

Authors would like to thank Patrick McKinlay for his work on the preliminary results for this paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Mathematical Appendix

Here, we provide all the formal proofs for the results in the manuscript.
Lemma A1.
Let ϕ ˜ ( X ) be the fitness function and let C n * x L n . Then Φ x and Φ y can be simplified to
Φ x ( C n ) = L + i = 1 n 1 a ˜ y i ( Δ x i ) + b ˜ x i + 1 ( Δ y i ) + 1 2 y i ( Δ x i 2 ) + 1 2 x i + 1 ( Δ y i 2 )
and
Φ y ( C n ) = L + i = 1 n 1 a ˜ y i + 1 ( Δ x i ) + b ˜ x i ( Δ y i ) + 1 2 y i + 1 ( Δ x i 2 ) + 1 2 x i ( Δ y i 2 ) ,
where L : = 2 a ˜ b ˜ + 1 2 ( a ˜ + b ˜ ) .

Appendix A.1. Proof of Theorem 1

Proof. 
Let a ˜ = b ˜ and let { x k * x , y k * x } k = 1 n L n be the optimal strategy in the x direction. Since
i = 1 n 1 y i ( Δ x i ) + x i + 1 ( Δ y i ) = i = 1 n 1 y i x i + x i + 1 y i + 1 = i = 1 n 1 Δ ( y i x i ) = y n x n x 1 y 1 = 1 .
we obtain in addition to Lemma A1
Φ x ( x i , y i ) = 2 a ˜ 2 + 2 a ˜ + 1 2 i = 1 n 1 y i ( Δ ( x i ) 2 ) + x i + 1 ( Δ ( y i ) 2 ) = 2 a ˜ 2 + 2 a ˜ + 1 2 i = 1 n 1 x i + 1 ( Δ ( y i ) 2 ) + y i ( Δ ( x i ) 2 ) = Φ y ( y i , x i ) .
Hence, if C n * x = { ( x i * x , y i * x ) 1 i n } maximises Φ x , then C n * y = { ( y i * x , x i * x ) 1 i n } maximises Φ y . This also holds in the opposite direction and completes the proof. □

Appendix A.2. Proof of Proposition 1

Proof. 
If C n * optimises Φ x , then it necessarily solves
x k Φ x ( C n * ) = x k * ( Δ y k 1 * ) + 1 2 ( Δ y k 1 * ) ( y k * + y k 1 * ) = 0 , y k Φ x ( C n * ) = 1 2 ( Δ x k * ) ( x k * + x k + 1 * ) ( Δ x k * ) y k * = 0
for k = 2 , , n 1 . By conditions (a)–(d), Δ x k * , Δ y k * 0 , requiring
2 x k * = y k * + y k 1 *
2 y k * = x k * + x k + 1 *
for k = 2 , , n 1 . Consequently, for 2 k n 2 ,
2 x k + 1 * = ( A3 ) y k * + y k + 1 * = ( A4 ) 1 2 ( x k * + x k + 1 * ) + 1 2 ( x k + 1 * + x k + 2 * ) ,
that is,
Δ 2 x k * = x k + 2 * 2 x k + 1 * + x k * = 0 .
Therefore, Δ x k * is constant for k = 2 , , n 1 , implying Δ x k * = 1 x 2 * n 2 = : C for k = 2 , , n 1 . Hence,
x k * = x 2 * + ( k 2 ) C = ( k 2 ) + x 2 * ( n k ) n 2 , k = 2 , , n .
Similarly, we obtain Δ 2 y k * = 0 for k = 1 , , n 3 and therefore Δ y k * = y n 1 * n 2 = : D for k = 1 , , n 2 . This implies that
y k * = y n 1 * ( n 1 k ) D = y n 1 * ( k 1 ) n 2 , k = 1 , , n 1 .
In order to obtain x 2 * and y n 1 * , note that (A3) implies x 2 * = 1 2 y 2 * and (A4) y n 1 * = 1 2 ( 1 + x n 1 * ) . Using the solutions, we have
x 2 * = 1 2 y 2 * = y n 1 * 2 ( n 2 )
and therefore
y n 1 * = 1 2 ( 1 + x n 1 * ) = 1 2 2 n 5 n 2 + y n 1 * 2 ( n 2 ) 2 .
Solving for y n 1 * leads to
y n 1 * = 2 ( n 2 ) 2 n 3 ,
which implies
x 2 * = y n 1 * 2 ( n 2 ) = 1 2 n 3 .
Clearly 0 < x 2 * < 1 for n 3 by the same argument as 0 < y n 1 * < 1 . We therefore obtain a unique increasing sequence in x * and y * that satisfies the necessary condition to maximise Φ x given by
x k * = 0 k = 1 2 k 3 2 n 3 k = 2 , , n 1 1 k = n , y k * = 0 k = 1 2 k 2 2 n 3 k = 2 , , n 1 1 k = n .
In order to check whether the critical point is a maximum or a minimum, we construct the Hessian matrix, which has the block form
H = D 1 B B T D 2 ,
then its determinant is given by
det ( H ) = det ( D 1 D 2 B B T ) ,
where, at the equilibrium, we have
D 1 = diag ( Δ y 1 , , Δ y n 2 ) = c I , D 2 = diag ( Δ x 2 , , Δ x n 1 ) = c I ,
where c = 2 2 n 3 and B is a bi-diagonal matrix with diagonal entries given by ( y 2 x 2 , y 3 x 3 , , y n 1 x n 1 ) and subdiagonal entries given by ( x 3 y 2 , x 4 y 3 , , x n 1 y n 2 ) . Then,
det ( H ) = det ( c 2 I B B T ) .
The matrix D 1 D 2 B B T is of the same structure as B B T (Tridiagonal/Band matrix). The n 2 eigenvalues of B are given by its diagonal entries
ν B i = y i x i = 1 ( 2 n 3 ) .
Hence, B has one eigenvalue ν B = 1 ( 2 n 3 ) with multiplicity n 2 . Applying the singular value decomposition of B, we obtain eigenvalues of B B T , which is ν B B T = 1 ( 2 n 3 ) 2 with multiplicity n 2 .
Let ν H be an eigenvalue of H, then
0 = det ( H ν H I 2 ( n 2 ) ) = det D 1 ν H I n 2 B B T D 2 ν H I n 2 = ( A6 ) det ( ( c I n 2 ν H I n 2 ) 2 B B T ) = det ( ( c ν H ) 2 ψ I n 2 B B T ) ,
and therefore ψ is an eigenvalues of B B T . Hence, the eigenvalues of H are given by ν H = c ± ν B B T and we obtain
ν H = 2 2 n 3 ± 1 2 n 3 2 2 n 3 + 1 2 n 3 < 0 .
This implies that H evaluated at the proposed optimal path is negative definite. Then, the path C n * = { x i * x , y i * x : 1 i n } maximises Φ x . By Theorem 1, { x i * y , y i * y } i = 1 n = { y i * x , x i * x } i = 1 n is the optimal solution in the y direction, which completes the proof. □

Appendix A.3. Proof of Proposition 2

Proof. 
As before, we maximise Φ x over the learning path. Note that further we shall omit the super-script x . Taking the partial derivatives yields for 2 k n 1
x k Φ x = L e m 1 ( Δ y k 1 ) δ x k + 1 2 ( y k 1 + y k ) = 0 , y k Φ x = L e m 1 ( Δ x k ) δ + 1 2 ( x k + x k + 1 ) y k = 0 .
By properties (a)–(b), that is, Δ x k 0 and Δ y k 0 , k = 1 , , n 1 , we obtain the necessary conditions for the critical points
x k * = δ + 1 2 ( y k * + y k 1 * ) , 2 k n 1
y k * = δ + 1 2 ( x k * + x k + 1 * ) , 2 k n 1 .
Together, we obtain for 3 k n 1
x k * = ( A7 ) δ + 1 2 ( y k * + y k 1 * ) = ( A8 ) 1 2 x k * + 1 4 x k + 1 * + 1 4 x k 1 *
or simply
0 = 2 x k * + x k + 1 * + x k 1 * = Δ 2 x k 1 * , 3 k n 1 .
This implies that Δ x k * : = c R for 2 k n 1 . Then, the interval [ x 2 * , 1 ] is divided into n 2 equal sized segments. Consequently, c = 1 x 2 * n 2 and therefore for 2 k n
x k * = x 2 * + ( k 2 ) c = x 2 * + ( k 2 ) 1 x 2 * n 2 = n k n 2 x 2 * + k 2 n 2 .
Similarly, we obtain Δ 2 y k * = 0 , k = 1 , , n 3 . Hence, Δ y k * : = d R for 1 k n 2 , implying that the interval [ 0 , y n 1 * ] is distributed into n 2 equally sized segments, such that d = y n 1 * n 2 . Then,
y k * = ( k 1 ) d = ( k 1 ) y n 1 * n 2 , 1 k n 1 .
To find x 2 * and y n 1 * , we utilise relations (A7) and (A8) as follows
x 2 * = ( A7 ) δ + 1 2 y 2 * = ( A10 ) δ + y n 1 * 2 ( n 2 )
and
y n 1 * = ( A8 ) δ + 1 2 x n 1 * + 1 2 = ( A9 ) δ + 1 2 + 1 2 x 2 * + ( n 3 ) 1 x 2 * n 2 .
Combining both equations yields
x 2 * = δ + 1 + δ 2 n 3 and y n 1 * = ( 1 + δ ) 2 ( n 2 ) 2 n 3 ,
which is only feasible for δ < 1 2 ( n 2 ) to guarantee x 2 * , y n 1 * ( 0 , 1 ] , or if δ < 1 , then it is only feasible for n = 2 . Hence, the only path satisfying the necessary condition to be the optimal path in the x direction is given by
x k * = 0 k = 1 δ + ( 1 + δ ) 2 k 3 2 n 3 2 k n , y k * = ( 1 + δ ) 2 k 2 2 n 3 1 k n 1 1 k = n .
To support the claim that this learning path maximises the fitness-over-learning function we consider the Hessian matrix for Φ x
H = D 1 B B T D 2 ,
where D 1 and D 2 are diagonal matrices of the form
D 1 = diag ( Δ y 1 , Δ y 2 , , Δ y n 2 ) ,
D 2 = diag ( Δ x 2 , Δ x 3 , , Δ x n 1 ) ,
and B being a bi-diagonal matrix with diagonal entries given by
( δ + ( y 2 x 2 ) , δ + ( y 3 x 3 ) , , δ + ( y n 1 x n 1 ) ) ,
and subdiagonal entries
( δ + ( x 3 y 2 ) , δ + ( x 4 y 3 ) , , δ + ( x n 1 y n 2 ) ) .
As in Proposition 1 we obtain
det ( H ) = det ( D 1 D 2 B B T )
and the eigenvalues of H, ν H , satisfy
c + ν H 2 = v B B T = ν B 2 v H = c ± | v B | ,
where ν B are the eigenvalues of B given by the diagonal entries, namely, ν B i = δ + ( y i * x i * ) and c = Δ x i * = Δ y i 1 * = 2 ( 1 + δ ) ( 2 n 3 ) . For a feasible solution, c > 0 . Hence the eigenvalues of H are negative as
| δ + ( y i * x i * ) | = | 1 + δ | 2 n 3 < 2 | 1 + δ | ( 2 n 3 ) .
Therefore, | v B | < c and v H < 0 . This completes the proof that the learning path is optimal for the x direction. □

Appendix A.4. Proof of Proposition 3

Proof. 
The partial derivatives of Φ y are
y k Φ y = L e m 1 y k 2 a ˜ b ˜ + 1 2 ( a ˜ + b ˜ ) + i = 1 n 1 a ˜ y i + 1 ( Δ x i ) + b ˜ x i ( Δ y i ) + 1 2 y i + 1 ( Δ x i 2 ) + 1 2 x i ( Δ y i 2 ) = a ˜ ( Δ x k 1 ) + b ˜ x k 1 b ˜ x k + 1 2 ( Δ x k 1 2 ) + 1 2 x k 1 2 y k 1 2 x k 2 y k = ( Δ x k 1 ) δ + 1 2 ( x k + x k 1 ) y k , x k Φ y = L e m 1 x k 2 a ˜ b ˜ + 1 2 ( a ˜ + b ˜ ) + i = 1 n 1 a ˜ y i + 1 ( Δ x i ) + b ˜ x i ( Δ y i ) + 1 2 y i + 1 ( Δ x i 2 ) + 1 2 x i ( Δ y i 2 ) = a ˜ y k a ˜ y k + 1 + b ˜ ( Δ y k ) + 1 2 y k 2 x k 1 2 y k + 1 2 x k + 1 2 ( Δ y k 2 ) = ( Δ y k ) δ x k + 1 2 ( y k + 1 + y k ) .
By properties (a)–(b), that is, Δ x k , Δ y k 0 for k = 1 , , n 1 , we obtain the necessary conditions
δ + 1 2 ( x k + x k 1 ) y k = 0 , 2 k n 1 δ x k + 1 2 ( y k + 1 + y k ) = 0 , 2 k n 1 ,
which can be rewritten as
x k * = δ + 1 2 ( y k + 1 + y k ) , 2 k n 1
y k * = δ + 1 2 ( x k + x k 1 ) , 2 k n 1 .
In the similar manner as in the proof of Proposition 2, we obtain
Δ y k * = 1 y 2 * n 2 for 2 k n 1 and Δ x k * = x n 1 * n 2 for 1 k n 2 ,
and finally the optimal solution of the form
y k * = 0 , k = 1 δ + ( 1 δ ) 2 k 3 2 n 3 , k = 2 , , n 1 1 , k = n , x k * = 0 , k = 1 ( 1 δ ) 2 k 2 2 n 3 , k = 2 , , n 1 1 , k = n .
This is feasible only if δ > 1 2 ( n 2 ) to guarantee y k * , x k * 0 for 1 k n . In the similar manner to Proposition 2 one can show that it is, indeed, a maximum of Φ y . □

Appendix A.5. Proof of Theorem 2

Proof. 
We use the fitness-over-learning functions at the equilibrium and obtain
Φ x * Φ y * = δ × 3 δ 2 ( n 1 ) ( n 2 ) + 1 ( 2 n 3 ) 2 .
Since 3 δ 2 ( n 1 ) ( n 2 ) + 1 ( 2 n 3 ) 2 > 0 , for δ > 0 the y direction is optimal and for δ < 0 it is the x direction. □

Appendix A.6. Proof of Corollary 1

Proof. 
The statement follows from the fact that for δ 1 , the optimal learning path of order n in the y direction is only feasible for n = 2 , resulting in a one-step function. Since the optimal path in the x direction is unfeasible for such δ for all n, the y direction is chosen first. □

Appendix A.7. Proof of Corollary 2

Proof. 
The optimal solution in the y direction satisfies the condition:
y k * , y = δ + ( 1 δ ) ( 2 k 3 ) 2 n 3 = δ + ( 1 δ ) ( 2 k 2 ) 2 n 3 ( 1 δ ) 2 n 3 = δ + x k * , y 1 δ 2 n 3 .
Hence,
y * , y : = lim n y k * , y = lim n ( δ + x k * , y ) ,
or, since as n , then k , we simply obtain
y * , y = δ + x * , y , x * , y ( 0 , 1 δ ) .
Boundary values for x * , y = 0 and x * , y ( 1 δ , 1 ] follow from the constraints 0 y * , y 1 . □

References

  1. Broom, M.; Jan, R. Game-Theoretical Models in Biology; CRC Press, Taylor & Francis Group: Boca Raton, FL, USA, 2013. [Google Scholar]
  2. Weibull, J.W. Evolutionary Game Theory; MIT Press: Cambridge, MA, USA, 1997. [Google Scholar]
  3. Fudenberg, D.; Levine, D.K. The Theory of Learning in Games; The MIT Press: Cambridge, MA, USA, 1999. [Google Scholar]
  4. Fudenberg, D.; Levine, D. Learning in games. Eur. Econ. Rev. 1998, 42, 631–639. [Google Scholar] [CrossRef]
  5. Selten, R. Reexamination of the perfectness concept for equilibrium points in extensive games. Int. J. Game Theory 1975, 4, 25–55. [Google Scholar] [CrossRef] [Green Version]
  6. Stadler, P.F.; Schuster, P. Mutation in Autocatalytic Reaction Networks. J. Math. Biol. 1992, 30, 597–631. [Google Scholar] [CrossRef]
  7. Tarnita, C.E.; Antal, T.; Nowak, M.A. Mutation–selection equilibrium in games with mixed strategies. J. Theor. Biol. 2009, 261, 50–57. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Wu, B.; Gokhale, C.S.; Wang, L.; Traulsen, A. How small are small mutation rates? J. Math. Biol. 2012, 64, 803–827. [Google Scholar] [CrossRef] [PubMed]
  9. Archibald, C.; Shoham, Y. Hustling in repeated zero-sum games with imperfect execution. In Proceedings of the 22nd International Joint Conference on Artificial Intelligence, Barcelona, Spain, 16–22 July 2011. [Google Scholar]
  10. Bowling, M.; Veloso, M. Existence of multiagent equilibria with limited agents. J. Artif. Intell. Res. 2004, 22, 353–384. [Google Scholar] [CrossRef]
  11. Borkar, V.S.; Jain, S.; Rangarajan, G. Evolutionary games with two timescales. Phys. D Nonlinear Phenom. 1999, 125, 155–166. [Google Scholar] [CrossRef]
  12. Beck, J.D.; Ejov, V.; Filar, J.A. Incompetence and impact of training in bimatrix games. Automatica 2012, 48, 2400–2408. [Google Scholar] [CrossRef]
  13. Beck, J.D. Incompetence, Training and Changing Capabilities in Game Theory. Ph.D. Thesis, University of South Australia, Adelaide, Australia, 2013. [Google Scholar]
  14. Kleshnina, M.; Filar, J.A.; Ejov, V.; McKerral, J.C. Evolutionary games under incompetence. J. Math. Biol. 2018, 77, 627–646. [Google Scholar] [CrossRef] [Green Version]
  15. Smith, J.M. Evolution and the Theory of Games; Cambridge University Press: Cambridge, UK, 1982. [Google Scholar]
  16. May, R.M. More evolution of cooperation. Nature 1987, 327, 15–17. [Google Scholar] [CrossRef]
  17. Sigmund, K.; Silva, H.D.; Hauert, C.; Traulsen, A. Social learning promotes institutions for governing the commons. Nature 2010, 466, 861. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Nowak, M.A. Evolutionary Dynamics: Exploring the Equations of Life; The Belknap Press of Harvard University Press: Cambridge, MA, USA, 2006. [Google Scholar]
  19. Ågren, J.A.; Davies, N.G.; Foster, K.R. Enforcement is central to the evolution of cooperation. Nat. Ecol. Evol. 2019, 3, 1018–1029. [Google Scholar] [CrossRef]
  20. Axelrod, R. The emergence of cooperation among egoists. Am. Political Sci. Rev. 1981, 75, 306–318. [Google Scholar] [CrossRef] [Green Version]
  21. Axelrod, R.; Hamilton, W.D. The evolution of cooperation. Science 1981, 211, 1390–1396. [Google Scholar] [CrossRef]
  22. Doebeli, M.; Hauert, C. Models of cooperation based on the Prisoner’s Dilemma and the Snowdrift game. Ecol. Lett. 2005, 8, 748–766. [Google Scholar] [CrossRef] [Green Version]
  23. Chao, L.; Elena, S.F. Nonlinear trade-offs allow the cooperation game to evolve from Prisoner’s Dilemma to Snowdrift. Proc. R. Soc. B Biol. Sci. 2017, 284, 20170228. [Google Scholar] [CrossRef] [Green Version]
  24. Anh, H.T.; Pereira, L.M.; Santos, F.C. Intention recognition promotes the emergence of cooperation. Adapt. Behav. 2011, 19, 264–279. [Google Scholar]
  25. Segbroeck, S.V.; Jong, S.D.; Nowé, A.; Santos, F.C.; Lenaerts, T. Learning to coordinate in complex networks. Adapt. Behav. 2010, 18, 416–427. [Google Scholar] [CrossRef] [Green Version]
  26. Hilbe, C.; Schmid, L.; Tkadlec, J.; Chatterjee, K.; Nowak, M.A. Indirect reciprocity with private, noisy, and incomplete information. Proc. Natl. Acad. Sci. USA 2018, 115, 12241–12246. [Google Scholar] [CrossRef] [Green Version]
  27. Draghi, J.; Whitlock, M.C. Overdominance interacts with linkage to determine the rate of adaptation to a new optimum. J. Evol. Biol. 2015, 28, 95–104. [Google Scholar] [CrossRef]
  28. Hofbauer, J.; Sigmund, K. Evolutionary game dynamics. Bull. Am. Math. Soc. 2003, 40, 479–519. [Google Scholar] [CrossRef] [Green Version]
  29. Taylor, P.D.; Jonker, L.B. Evolutionary stable strategies and Game Dynamics. Math. Biosci. 1978, 40, 145–156. [Google Scholar] [CrossRef]
  30. Hofbauer, J.; Schuster, P.; Sigmund, K.; Wolff, R. Dynamical Systems Under Constant Organization II: Homogeneous Growth Functions of Degree p = 2. SIAM J. Appl. Math. 1980, 38, 282–304. [Google Scholar] [CrossRef]
  31. Kleshnina, M. Evolutionary Games under Incompetence & Foraging Strategies of Marine Bacteria. Ph.D. Thesis, University of Queensland, Brisbane, Australia, 2019. [Google Scholar]
  32. Cressman, R.; Tao, Y. The replicator equation and other game dynamics. Proc. Natl. Acad. Sci. USA 2014, 111, 10810–10817. [Google Scholar] [CrossRef] [Green Version]
Figure 1. (A) A schematic representation of the effect of mistakes: individual 1 chooses to cooperate and individual 2 chooses to defect. However, under incompetence, individual 1 will cooperate only with probability α and individual 2 will defect only with probability β . Hence, the outcome of the interaction now depends on the probability distribution of mistakes. (B) A schematic representation of prioritised learning: we aim to define first an optimal direction in which the learning should start (x and y direction represents strategies 1 and 2, respectively) and an n-step optimal sequence of ( x k x , y k x ) if x direction is optimal (or ( x k y , y k y ) if y direction is optimal). (C) An example of a learning path in the x direction. We start at ( x 0 , y 0 ) , then taking the first half-step in the x direction to ( x 1 , y 0 ) and a second half-step in the y direction to ( x 1 , y 1 ) . By design, we have to start at ( 0 , 0 ) and finish at ( 1 , 1 ) .
Figure 1. (A) A schematic representation of the effect of mistakes: individual 1 chooses to cooperate and individual 2 chooses to defect. However, under incompetence, individual 1 will cooperate only with probability α and individual 2 will defect only with probability β . Hence, the outcome of the interaction now depends on the probability distribution of mistakes. (B) A schematic representation of prioritised learning: we aim to define first an optimal direction in which the learning should start (x and y direction represents strategies 1 and 2, respectively) and an n-step optimal sequence of ( x k x , y k x ) if x direction is optimal (or ( x k y , y k y ) if y direction is optimal). (C) An example of a learning path in the x direction. We start at ( x 0 , y 0 ) , then taking the first half-step in the x direction to ( x 1 , y 0 ) and a second half-step in the y direction to ( x 1 , y 1 ) . By design, we have to start at ( 0 , 0 ) and finish at ( 1 , 1 ) .
Mathematics 08 01945 g001
Figure 2. Optimal learning paths for examples in the manuscript. (A) Examples 1 and 2: optimal learning paths for R 1 and R 2 follow the line x = y (blue line). We plot the solution (red stepwise line) for: (1) n = 5 , (2) n = 10 . (B) Example 3: an optimal learning path for R 3 where n = 12 and: (1) a = 5 , (2) a = 7 .
Figure 2. Optimal learning paths for examples in the manuscript. (A) Examples 1 and 2: optimal learning paths for R 1 and R 2 follow the line x = y (blue line). We plot the solution (red stepwise line) for: (1) n = 5 , (2) n = 10 . (B) Example 3: an optimal learning path for R 3 where n = 12 and: (1) a = 5 , (2) a = 7 .
Mathematics 08 01945 g002
Figure 3. The fitness-over-learning Φ y * calculated for the optimal solution from (14) for different values of δ > 0 and different step size n. (A) The colormap represents dependence of Φ y * on the number of steps and δ . (B) We plot the exact values of Φ y * for three values of δ ( 0 , 1 ) . As can be seen, the increasing number of steps improves the values of the fitness-over-learning function. (C) However, applying the same solution for higher values of the relative strategic advantage ( δ > 1 ) reduces the value of the objective function, see Corollary 1.
Figure 3. The fitness-over-learning Φ y * calculated for the optimal solution from (14) for different values of δ > 0 and different step size n. (A) The colormap represents dependence of Φ y * on the number of steps and δ . (B) We plot the exact values of Φ y * for three values of δ ( 0 , 1 ) . As can be seen, the increasing number of steps improves the values of the fitness-over-learning function. (C) However, applying the same solution for higher values of the relative strategic advantage ( δ > 1 ) reduces the value of the objective function, see Corollary 1.
Mathematics 08 01945 g003
Table 1. Definitions of advantages of cooperators over defectors. For the definition of advantages of defectors over cooperators, the inequality signs in parameters comparison should be reversed.
Table 1. Definitions of advantages of cooperators over defectors. For the definition of advantages of defectors over cooperators, the inequality signs in parameters comparison should be reversed.
AdvantageParameters ComparisonDiscussion
Fitness advantage a > b Fitness advantage implies that cooperator have higher fitness and, hence, are more abundant.
Learning advantage α < β Learning advantage implies that cooperators are more flexible in their strategy execution and can act as defectors.
Strategic advantage δ > 0 or a ˜ > b ˜ Strategic advantage combines both fitness and learning advantages implying that if cooperators are disadvantaged in fitness (or learning), then advantage in learning (or fitness) can compensate.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kleshnina, M.; Streipert, S.S.; Filar, J.A.; Chatterjee, K. Prioritised Learning in Snowdrift-Type Games. Mathematics 2020, 8, 1945. https://doi.org/10.3390/math8111945

AMA Style

Kleshnina M, Streipert SS, Filar JA, Chatterjee K. Prioritised Learning in Snowdrift-Type Games. Mathematics. 2020; 8(11):1945. https://doi.org/10.3390/math8111945

Chicago/Turabian Style

Kleshnina, Maria, Sabrina S. Streipert, Jerzy A. Filar, and Krishnendu Chatterjee. 2020. "Prioritised Learning in Snowdrift-Type Games" Mathematics 8, no. 11: 1945. https://doi.org/10.3390/math8111945

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop