Next Article in Journal
An Agent-Based Model of Institutional Life-Cycles
Previous Article in Journal
The Seawall Bargaining Game
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning in Networks—An Experimental Study Using Stationary Concepts

1
Institute of Economics, Karlsruhe Institute of Technology (KIT), P.O. Box 6980, Karlsruhe D-76049, Germany
2
Otto-von-Guericke University Magdeburg, Faculty of Economics and Management, P.O. Box 4120, Magdeburg 39016, Germany
*
Author to whom correspondence should be addressed.
Games 2014, 5(3), 140-159; https://doi.org/10.3390/g5030140
Submission received: 21 January 2014 / Revised: 8 July 2014 / Accepted: 9 July 2014 / Published: 31 July 2014
(This article belongs to the Special Issue Social Networks and Network Formation 2013)

Abstract

:
Our study analyzes theories of learning for strategic interactions in networks. Participants played two of the 2 × 2 games used by Selten and Chmura [1]. Every participant played against four neighbors. As a distinct aspect our experimental design allows players to choose different strategies against each different neighbor. The games were played in two network structures: a lattice and a circle. We analyze our results with respect to three aspects. We first compare our results with the predictions of five different equilibrium concepts (Nash equilibrium, quantal response equilibrium, action-sampling equilibrium, payoff-sampling equilibrium, and impulse balance equilibrium) which represent the long-run equilibrium of a learning process. Secondly, we relate our results to four different learning models (impulse-matching learning, action-sampling learning, self-tuning EWA, and reinforcement learning) which are based on the (behavioral) round-by-round learning process. At last, we compare the data with the experimental results of Selten and Chmura [1]. One main result is that the majority of players choose the same strategy against each neighbor. As other results, we observe an order of predictive success for the equilibrium concepts that is different from the order shown by Selten and Chmura and an order of predictive success for the learning models that is only slightly different from the order shown in a recent paper by Chmura, Goerg and Selten [2].

1. Introduction

In their 2008 paper, Reinhard Selten and Thorsten Chmura (henceforth SC) [1] analyze a set of 12 different 2 × 2 games. For 6 constant and 6 non-constant sum games they compare the predictive success of five stationary concepts. The five concepts compared are: (1) Nash equilibrium (Nash); (2) Quantal response equilibrium (QRE) [3]; (3) Action-sampling equilibrium (ASE) [1]; (4) Payoff-sampling equilibrium (PSE) [4]; and (5) Impulse balance equilibrium (IBE) [5,6]. Since these concepts are explained in detail in SC [1], we do not explain them here. In their study, the randomly matched participants played the games over 200 rounds, which allows an interpretation of the concepts “... as stationary states of dynamic learning models” ([1], p. 940). In our study we put the focus on learning in networks.
We recall a point made often and in many situations, that learning occurs in social contexts. When analyzing learning in economic decision-making, most studies deal with repeated interactions between two or more randomly matched players. In real life, learning in networks seems to be more natural. More precisely, in many applications, local interaction in which most of the players interact most of the time with only a subset of the population seems to be a more appropriate approach. Therefore, we analyze whether or not there are differences in behavior (or learning) in networks compared to behavior (learning) in an environment of random matching.
We first analyze our experimental results in relation to the long-run equilibrium of the learning process, predicted by the five different equilibrium concepts to check for such differences. We relate our results to the predictions of different learning models tested in a recent paper of Chmura et al. (henceforth CGS) [2] to shed light on the learning process itself. The four learning models used are: (1) Action-sampling learning (ASL) [2]; (2) Impulse-matching learning (IML) [2]; (3) Self-tuning experience-weighted attraction learning (EWA) [7]; and (4) Reinforcement learning (REL) [8]. Since one can find a detailed description of the learning models we use in CGS [2], we do not explain them here.
We change the neighborhood from random matching to a network structure to test the impact of this parameter. As a distinctive aspect our experimental design is such that players are allowed to choose different strategies with their different neighbors. From this it follows that every pair of linked players can play each game as an isolated game. This distinguishes our design from most other studies of games on networks, where players are forced to choose the same strategy with their neighbors. Thus, the question of whether players actually choose different strategies against their network partners is another topic which is investigated in this study.
While in a random matching environment players can mix strategies over time, in our network they can additionally mix strategies within one period by playing different strategies against their neighbors. Learning in network structures also occurs via indirect neighbors whose decisions also affect direct neighbors.1
In our study, we design a neighborhood game in an exogenously fixed network, where the players cannot choose their neighbors but can choose different strategies against each of the exogenously given neighbors. We run two different games used by SC as baseline games in a network. The experimental results we present allow us to analyze on the one hand how network structures affect learning, and on the other hand how to control for the predictive success of different equilibrium concepts and of different models of learning. Our analysis is guided by four key questions:
  • Key question 1. Do the participants (actually) mix strategies?
  • Key question 2. Does the structure of the network affect learning in the games?
  • Key question 3. Are the equilibrium concepts tested by Selten and Chmura [1] good predictors for the participants’ behavior in network games?
  • Key question 4. Are the learning models tested by Chmura et al. [2] good predictors for the participants’ learning process in network games?2
Based on the games used by SC (a constant and a non-constant sum game), we construct neighborhood games, where each player has four direct neighbors. With respect to the idea that players have to decide how they interact with their partners and how they adjust their behavior over time, the participants in our experiment could choose different strategies for each neighbor. In contrast to most other studies on behavior in networks, participants in our framework have the opportunity to mix strategies both inter-temporally and intra-temporally.
Guided by the findings of Berninghaus et al. [12], where the players’ behavior was affected by different network structures, we used two different structures: a lattice and a circle.
We address our first key question by analyzing how often the players use the ability provided, to play different strategies against their neighbors (to mix strategies within one round).
Comparing the participants’ behavior in the two different network structures, we answer our second key question. Additionally, comparing our observed results with those of SC, we want to provide evidence that learning in networks differs from learning in a random matching environment.
To answer our third key question, we compare our experimental results with the predictions of the five learning concepts. The results of SC were revised in studies by Brunner et al. [13] (henceforth BCG) and by Selten et al. [14]. To ensure that the results are comparable, we use the same statistical techniques as these studies to analyze our data.
As benchmarks for our analyses we use the mean frequencies for each model of learning reported by CGS, the mean frequencies predicted by the equilibrium concepts, and the observed frequencies in the experiment reported by SC. In the study of Chmura et al. [2], the mean frequencies for the models of learning result from the run of simulations.3
As one of our main results we show that learning in networks is different from learning in random matching. However, we find no significant difference between the two network structures. In line with this result, we observe an order of predictive success for the five equilibrium concepts that differs from the order given by SC, which means that learning in networks has a significant impact on the long-run equilibrium, which occurs in the process of learning in networks. This result is supported by our finding regarding the predictions of the four different models of learning. While CGS shows that self-tuning EWA outperformed the other models of learning in random matching, we observe that action-sampling learning best predicts behavior in the network games we used.
As another remarkable result, we find that the majority of players choose the same strategy against each neighbor, i.e., players do not really mix strategies. This holds for both network structures employed in our experiments. Moreover, the average number of players not mixing strategies is slightly higher in the circle network.

2. Description of the Experiment

In our study we use two games (see Figure 1) taken from SC. The main reason we use these games is that they generate clearly distinct predictions for the five equilibrium concepts. Figure 2 shows the theoretical predictions of these concepts for the constant sum game and Figure 3 for the non-constant sum game. Figure 4 (Figure 5) shows the predictions of the four learning models for the constant sum game (non-constant sum game) we relate our experimental results to.4
Figure 1. The constant and the non-constant sum game.
Figure 1. The constant and the non-constant sum game.
Games 05 00140 g001
Figure 2. Theoretical equilibria in the constant sum games.
Figure 2. Theoretical equilibria in the constant sum games.
Games 05 00140 g002
Figure 3. Theoretical equilibria in the non-constant sum games.
Figure 3. Theoretical equilibria in the non-constant sum games.
Games 05 00140 g003
Figure 4. Predictions of the learning models for the constant sum games.
Figure 4. Predictions of the learning models for the constant sum games.
Games 05 00140 g004
Figure 5. Predictions of the learning models for the non-constant sum games.
Figure 5. Predictions of the learning models for the non-constant sum games.
Games 05 00140 g005

2.1. The Baseline Games

We use two games, a constant and a non-constant sum game, as baseline games. As explained by SC, both games form a pair that is characterized by having the same best response structure.5
Because we want, on the one hand, to turn the focus to learning by analyzing the impact of different network structures, and on the other hand to refer to the results of SC and BCG, we took the games from SC and extended them to neighborhood games in different network structures.

2.2. The Neighborhood Games and Network Structures

Following Berninghaus et al. [12], we ran the games in exogenously given fixed network structures where each player has four direct neighbors. To accomplish this, 16 players were allocated into two different network structures. The 2 × 2 games described in Section 2.1 represent the baseline games. Based on these games, the neighborhood game was constructed such that each player interacts with her four direct (local) neighbors. Suitably to our first key question, in our experiment each player could use a different strategy against each of her neighbors in each round.
Since it is well known in the literature that network structures affect the players’ behavior in games (see, e.g., [12,15,16]) we used two different structures. Guided by the findings of Berninghaus et al. [12] regarding the behavior of players in a coordination game, we used two different network structures: a lattice and a circle. Figure 6 and Figure 7 provide schematic illustrations of both network structures.
Figure 6. Network structure 1–a lattice.
Figure 6. Network structure 1–a lattice.
Games 05 00140 g006
Figure 7. Network structure 2–a circle.
Figure 7. Network structure 2–a circle.
Games 05 00140 g007
It is obvious that in both structures each player interacts with four direct neighbors. The difference between the structures is given by the number of indirect neighbors on the level of indirect interactions.
There might be more equilibria in the network game as we present in Figure 2 and Figure 3. Since we want to compare the players’ behaviors in the two different network structures as well as with the experimental results of SC, we use the equilibria presented in Section 2.1 as benchmarks.

2.3. Exemplary Explanations of the Interaction Structures

In the Lattice (Figure 6), player 1 interacts directly with player 2 as her right, player 4 as her left, player 13 as her top, and player 5 as her bottom neighbor6. On the first level of indirect interactions, player 1 has six indirect neighbors (players 16, 14, 3, 8, 6, and 9). On the second level of indirect interactions, player 1 has four indirect neighbors (players 15, 7, 12, and 10), and one indirect neighbor (player 11) on the third level of indirect interactions.
In the Circle (Figure 7), player 1 interacts directly with player 8 and player 16 as her left neighbors and with player 2 and player 10 as her right neighbors. On the first level of indirect interactions player 1 has five indirect neighbors (players 9, 7, 15, 11, and 3). On the second level of indirect interactions, player 1 has four indirect neighbors (players 6, 14, 12, and 4), and she has two indirect neighbors (players 13 and 5) on the third level of indirect interactions.

3. Experimental Procedure

In order to compare our results with the results of SC, we designed our experiment to be as similar as possible to theirs.
We ran the experiments at the MaXLab, the experimental laboratory at the University of Magdeburg. For each neighborhood game, four sessions were conducted with 16 participants each. The participants were randomly matched on a position in the network. We told them that the matching could (but did not have to) change during the experiment—which did not, in fact, happen. We ran the games over 100 rounds. In each round the players could choose a different strategy against each of the four direct neighbors. After each round the players were informed about the payoffs from each 2 × 2 game, i.e., about the total payoff from the last round and about the accumulated payoffs over all rounds.7
The payoffs were presented by “points” which were converted into Euro at an exchange rate of four payoff points equaling 1 Eurocent. An experimental session lasted about 1 to 1.5 h and the average earnings of a participant were about 10 Euro. The 256 participants in our experiment were recruited using ORSEE software [17] from a pool of mostly students from various faculties. To program our experiments we used the experimental software z-Tree [18].8

4. Experimental Results

To compare our experimental results with the theoretical predictions of the five concepts, and also with the results of SC, in Table 1 we show the observed relative frequencies of playing Up (strategy “up” in the 2 × 2 games) and Left (strategy “left” in the 2 × 2 games), played in the two different network structures.
Table 1. The relative frequencies of playing Up and Left in the baseline games.
Table 1. The relative frequencies of playing Up and Left in the baseline games.
Constant Sum Game
Session 1Session 2Session 3Session 4AverageVariance
LatticeUp0.0320.0530.0610.0450.0480.000153
Left0.7840.7020.5870.6920.6910.006529
CircleUp0.0360.0540.0480.0590.0490.000098
Left0.7790.6830.7560.6410.7150.004092
Non-Constant Sum Game
Session 1Session 2Session 3Session 4AverageVariance
LatticeUp0.0620.1380.1430.1000.1110.001425
Left0.6770.7640.7950.7410.7440.002500
CircleUp0.1150.1070.1100.0510.0960.000901
Left0.7240.8360.8630.6720.7740.008223
Key question 1. Do the participants (actually) mix strategies?
Unlike in other studies, the participants in our experiment could choose different strategies against each opponent in each period. Thus, the participants were able to mix their strategies within one round. As shown in Table 2, the participants did not use this opportunity very frequently.
Table 2. Average No. of players choosing the same strategy against each neighbor over 100 rounds.
Table 2. Average No. of players choosing the same strategy against each neighbor over 100 rounds.
Average No. of Players Choosing the Same Strategy against Each Neighbor over 100 Rounds
LatticeCircle
Constant sum game81.03%85.69%
Non-constant sum game81.16%86.22%
The majority of players chose the same strategy for each neighbor. We do not find any difference in the strategy selection for the base games, but there are slight differences between the two different network structures. In terms of learning, or adjusting behavior, the results show that the frequency of choosing the same strategy against each neighbor increases over time (see Table 3).
Table 3. Average No. of players choosing the same strategy in the first fifty and in the second fifty rounds.
Table 3. Average No. of players choosing the same strategy in the first fifty and in the second fifty rounds.
Average No. of players choosing the same strategy against each neighbor over rounds 1‑50
LatticeCircle
Constant sum game78.34%83.06%
Non-constant sum game76.66%83.09%
Average No. of players choosing the same strategy against each neighbor over rounds 51‑100
LatticeCircle
Constant sum game83.72%88.31%
Non-constant sum game85.66%89.34%
Result 1. Participants did not mix strategies within one round.
Key question 2. Does the structure of the network affect learning in games?
To find differences in learning between the Lattice network and the Circle network, we first analyze strategy selection in the baseline game. Based on the relative frequencies of playing Up and Left (see Table 1), we find no significant differences between the two network structures (sign test for any level of significance).
We compare our observed results with the experimental results of SC to analyze if there are at least differences between networks as compared to random matching environments with respect to learning behavior. Table 4 shows the observed results.
Table 4. The observed averages of playing Up and Left in games.
Table 4. The observed averages of playing Up and Left in games.
Strategy Choices—Observed Averages in:
Lattice NetworkCircle NetworkNo Network (SC [1])
Constant Sum Game
Up0.0480.0490.079
Left0.6910.7510.690
Non-constant Sum Game
Up0.1110.0960.141
Left0.7440.7740.564
As one can see (Figure 8 and Figure 9), our observed averages for both network structures are different from the results in the 2 × 2 games with only one randomly matched partner. The data in Figure 9 show that this is especially true for the non-constant sum game. These findings imply that learning in networks actually is different from learning in the 2 × 2 random matching environment used by SC.
Figure 8. The observed averages of playing Up and Left in the constant sum game.
Figure 8. The observed averages of playing Up and Left in the constant sum game.
Games 05 00140 g008
Figure 9. The observed averages of playing Up and Left in the non-constant sum game.
Figure 9. The observed averages of playing Up and Left in the non-constant sum game.
Games 05 00140 g009
Result 2. We found no significant differences in the participants’ behavior in the two network structures.
Key question 3. Are the equilibrium concepts tested by Selten and Chmura [1] good predictors for the participants’ behavior in network games?
As we want to answer the question about the predictive success of the five equilibrium concepts in the context of learning in networks, we first illustrate the theoretical predictions of the five concepts and the experimental results (Figure 10 and Figure 11).
Figure 10. Theoretical predictions of the five equilibrium concepts and the experimental results for the constant sum game.
Figure 10. Theoretical predictions of the five equilibrium concepts and the experimental results for the constant sum game.
Games 05 00140 g010
Figure 11. Theoretical predictions of the five equilibrium concepts and the experimental results for the non-constant sum game.
Figure 11. Theoretical predictions of the five equilibrium concepts and the experimental results for the non-constant sum game.
Games 05 00140 g011
We use the corrected theoretical predictions of SC, according to BCG. Table 5 shows the corresponding numerical predictions of the five equilibrium concepts.
Table 5. The theoretical predictions of the five equilibrium concepts and the observed relative frequencies of Up and Left strategies.
Table 5. The theoretical predictions of the five equilibrium concepts and the observed relative frequencies of Up and Left strategies.
NashQREASEPSEIBEObserved Average of Selten and ChmuraOur Observed Average in the LatticeOur Observed Average in the Circle
Constant sum game
Up0.0910.0420.0900.0710.0680.0790.0480.049
Left0.9090.6370.7050.6430.5800.6900.6910.751
Non-constant sum game
Up0.0910.0420.0900.0600.1040.1410.1110.096
Left0.9090.6370.7050.6910.6340.5640.7440.774
To measure the predictive success of the equilibrium concepts, we analyze our data according to the method used by SC. The analysis is based on pairwise comparisons of the observed and the predicted relative frequencies. Table 6 shows the mean squared distances and the sampling variance for both base games played in the two network structures.
Table 6. Mean squared distances of the five equilibrium concepts.
Table 6. Mean squared distances of the five equilibrium concepts.
NashQREASEPSEIBESampling Variance
Constant sum gameLattice0.0542970.0079880.0069860.0078800.0177980.005011
Circle0.0426190.0092410.0048990.0087640.0216520.003143
Non-const. sum gameLattice0.0304760.0191730.0049150.0083550.0151440.002943
Circle0.0251580.0284330.0116030.0149690.0264410.006843
Based on the mean squared distances, it can be seen that there is an ordering of the concepts in terms of success: action-sampling equilibrium, payoff-sampling equilibrium, quantal response equilibrium, impulse balance equilibrium, and Nash equilibrium. For the non-constant sum game, the impulse balance equilibrium performs slightly better than the quantal response equilibrium.
Following the analyses in SC and in BCG, we test the results of all 16 independent observations together as well as separately for the constant and the non-constant sum game. Since the theoretical predictions are independent of the network structure, we test the results for both structures together. As in SC, we run the Wilcoxon matched-pairs signed rank test to compare the squared distances of the five concepts from the observed relative frequencies. In Table 7, we show the p-values of a test checking whether the solution concept in the row outperforms the solution concept in the column. The bold numbers in the upper line are the p-values of a test using all 16 observations together. The middle line gives the p-values for the observations in the constant sum game, and the lower line the p-values for the observations in the non-constant game.
As per the remarks in BCG, we also perform the Kolmogorov-Smirnov two-sample test to double-check the significance of the results (p-values are the numbers in brackets given in Table 7).
Table 7. Predictive success-p-values in favor of the row concepts, above: 16 independent observations together, middle: constant sum game (eight independent observations), below: non-constant sum game (eight independent observations).
Table 7. Predictive success-p-values in favor of the row concepts, above: 16 independent observations together, middle: constant sum game (eight independent observations), below: non-constant sum game (eight independent observations).
PSEQREIBENash
ASEn. s.n. s.2% (10%)0.001% (0.01%)
n. s.n. s.10% (10%)0.02% (0.02%)
n. s.n. s.10% (n. s.)5% (n. s.)
PSE n. s.10% (n. s.)0.05% (2%)
n. s.n. s.0.1% (2%)
n. s.n. s.10% (n. s.)
QRE n. s.1% (5%)
n. s.0.2% (2%)
n. s.n. s.
IBE 5% (10%)
2% (10%)
n. s.
As per the remarks in BCG, we also perform the Kolmogorov-Smirnov two-sample test to double-check the significance of the results (p-values are the numbers in brackets given in Table 7).
When comparing all independent observations, it is obvious that all non-Nash concepts do significantly better than Nash.9
This holds for the constant sum game but not for the non‑constant sum game. Moreover, we found no significant clear order of predictive success among the four non-Nash concepts.
Result 3. There is no significant clear order of predictive success among the four non-Nash concepts.
Key question 4. Are the learning models tested by Chmura et al. [2] good predictors for the participants’ learning process in network games?
After analyzing the long-run equilibrium behavior in a learning process, we next concentrate on the round-by-round behavior and relate our observed results to the predictions of the four learning models. We first illustrate the predictions of the four learning models used (given by CGS) and the experimental results (Figure 12 and Figure 13).
Figure 12. Predictions of the four learning models and the experimental results for the constant sum game.
Figure 12. Predictions of the four learning models and the experimental results for the constant sum game.
Games 05 00140 g012
Figure 13. Predictions of the four learning models and the experimental results for the constant sum game.
Figure 13. Predictions of the four learning models and the experimental results for the constant sum game.
Games 05 00140 g013
Table 8 shows the corresponding numerical values of the predictions of the four learning models. We use the predictions given by CGS.
Table 8. The predictions of the four learning models and the observed relative frequencies of Up and Left strategies.
Table 8. The predictions of the four learning models and the observed relative frequencies of Up and Left strategies.
ASLIMLEWARELObserved Average of Selten and ChmuraOur Observed Average in the LatticeOur Observed Average in the Circle
Constant sum game
Up0.0950.0630.4990.1210.0790.0480.049
Left0.7100.5740.4990.3450.6900.6910.751
Non-constant sum game
Up0.0940.0980.4850.1700.1410.1110.096
Left0.7090.6370.5640.3800.5640.7440.774
Based on a pairwise comparison of the observed and the predicted relative frequencies we measure the predictive success of the four learning models. Table 9 shows the mean squared distances and the sampling variances.
Table 9. Mean squared distances of the four learning models.
Table 9. Mean squared distances of the four learning models.
IMLASLEWARELSampling Variance
Constant sum gameLattice0.0075960.0189920.1302660.2455980.005011
Circle0.0052590.0231430.1450060.2519660.003143
Non-const. sum gameLattice0.0044670.0146090.1391320.1754970.002943
Circle0.0110390.0255490.1673950.2023540.006843
Using the mean squared distances as a measure for predictive success, the following order can be derived: action-sampling learning, impulse-matching learning, reinforcement learning, and self-tuning EWA.
Following the analyses regarding the predictive success of the (long-run) equilibrium concepts, we use the Kolmogorov-Smirnov two-sample test to check for significance of the derived order of predictive success. We test the results for both structures together. Table 10 shows the p-values of the test of all 16 independent observations together as well as separately for the constant and the non-constant sum game. Table 10 is equally structured as Table 7, meaning we show the p-values of a test checking whether the learning model in the row outperforms the learning model in the column.
Table 10. Predictive success-p-values in favor of the row concepts, above: 16 independent observations together, middle: constant sum game (eight independent observations), below: non-constant sum game (eight independent observations).
Table 10. Predictive success-p-values in favor of the row concepts, above: 16 independent observations together, middle: constant sum game (eight independent observations), below: non-constant sum game (eight independent observations).
IMLRELEWA
ASL3.5%<0.1%<0.1%
8.7%<0.1%<0.1%
66.0%<0.1%<0.1%
IML <0.1%<0.1%
<0.1%<0.1%
<0.1%<0.1%
REL 0.3%
0.2%
8.7%
When comparing all 16 independent observations together, it is obvious that the results of the tests support our derived order of predictive success, i.e., action-sampling learning predicts the round-by-round behavior best. This holds for the constant sum game but not for the non-constant sum game. For the non-constant sum game the predictions of action-sampling learning and impulse-matching learning are not significantly different.
Result 4. There is a significant clear order of predictive success among the four learning models, which is: (1) action-sampling learning, (2) impulse-matching learning, (3) reinforcement learning, and (4) self-tuning EWA.

5. Conclusions

In this paper, we analyzed learning in networks. Therefore, we added network structures to the 2 × 2 games taken from Selten and Chmura [1]. Starting with two different games (a constant and a non‑constant sum game) we construct a neighborhood game with two different network structures (a lattice and a circle), where each player has four direct (local) neighbors. Unlike in other studies, the participants in our experiment could choose a different strategy against each neighbor.
As our study is related to that of SC, we compare our observed results to their experimental results, and to the theoretical predictions of five (long-run) equilibrium concepts (Nash-equilibrium, quantal response equilibrium, action-sampling equilibrium, payoff-sampling equilibrium, and impulse balance equilibrium). We additionally relate our results to the predictions of four learning models (action-sampling learning, impulse-matching learning, reinforcement learning, and self-tuning EWA), used by Chmura et al. [2], to shed some light on the learning process.
Guided by our key questions, we first analyze our data to check if the participants really play mixed strategies. Because the majority of players choose the same strategy against each neighbor, we conclude that the participants did not mix their strategies in a round.
In a second step, we consider the influence of different network structures on the behavior of the participants. The differences between our results and the experimental results of SC provide evidence that learning in network differs from learning in 2 × 2 games with only one randomly matched partner. Unlike in other studies, we find no significant difference in the behavior of the participants in the two network structures we use. We find only slight differences in the number of participants choosing the same strategy against each neighbor.
Concerning our third key question, we show that none of the five equilibrium concepts provides an exact prediction for the participants’ behavior in our experiment. We show that all non-Nash concepts outperform the Nash concept. The order of predictive success of the four non-Nash concepts is different for the constant sum and the non-constant sum game. It is perhaps notable that action-sampling equilibrium and payoff-sampling equilibrium strategies do better than impulse balance equilibrium when both games are combined. Aside from this, there is no clear ranking among these four non-Nash concepts.
With respect to our forth key question, we show that none of the learning models exactly predicts the round-by-round behavior of the participants in our experiment. It becomes obvious, that there is a clear order of predictive success among the four learning models, with action-sampling learning as best and self-tuning EWA as least prediction model. This order is derived when comparing the data of both games together. While this order of predictive success holds for the constant sum game, it does not hold for the non-constant sum game. Here, action-sampling learning predicts as good as impulse-matching learning. However, our derived order of predictive success supports the general findings of CSG using aggregate frequencies.
For future research it looks promising to further develop theories and test different network structures in addition to further replications.

Appendix

Written Instructions (Translation)

Welcome to our today’s experiment! Below, you can find the description of the experiment and then you are asked to make a series of decisions. Please read the following information very carefully. If you have any questions, please ask before you start the experiment. Please note that during the whole experiment, communication with the other participants is not allowed. Thank you!
In total, 16 players take part in this experiment. Each participant plays either as a Player 1 type or as a Player 2 type. They maintain that role throughout the duration of the experiment.
You play the games over 100 rounds.
In every round, each Player 1 type plays with four Player 2 types and each Player 2 type plays with four Player 1 types. The combination of players can randomly change (but did not have to) in every round.
On the screen you see four matrices (for each pair of players, a matrix) with four fields. As the screenshot show, in each round and in each of the four matrices, you can decide between the two possible strategies (Strategy A or Strategy B).
Games 05 00140 i001
The payoff of the row player can be found in the upper left corner and of the column player in the lower right corner. Your (possible) payoffs are displayed with a round frame. Your payoff depends on your own choice and the choices of the other players. After making your choice, your selected line turns red. After the selection of the other players, in each matrix the field with the payoff you receive turns yellow.
After each round you will be notified which payoff (for each pairing/matrix) you received in the last round. In addition, you are informed about the total payoff (cumulated over all played matrices in the already played rounds. The conversion rate for your received payoffs will be announced on the screen.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Selten, R.; Chmura, T. Stationary concepts for experimental 2 × 2-games. Am. Econ. Rev. 2008, 98, 938–966. [Google Scholar] [CrossRef]
  2. Chmura, T.; Goerg, S.J.; Selten, R. Learning in experimental 2 × 2 games. Games Econ. Behav. 2012, 76, 44–73. [Google Scholar] [CrossRef]
  3. McKelvey, R.D.; Palfrey, T.R. Quantal response equilibria for normal form games. Games Econ. Behav. 1995, 10, 6–38. [Google Scholar] [CrossRef]
  4. Osborne, M.J.; Rubinstein, A. Games with procedurally rational players. Am. Econ. Rev. 1998, 88, 834–847. [Google Scholar]
  5. Selten, R.; Buchta, J. Experimental sealed bid first price auctions with directly observed bid functions. In Games and Human Behavior: Essays in the Honor of Amnon Rapoport; Budescu, D.V., Erev, I., Zwick, R., Eds.; Lawrence Erlbaum Associates: Mahwah, NJ, USA, 1999; pp. 79–104. [Google Scholar]
  6. Selten, R.; Abbink, K.; Cox, R. Learning direction theory and the winner’s curse. Exp. Econ. 2005, 8, 5–20. [Google Scholar] [CrossRef]
  7. Ho, T.; Camerer, C.F.; Chong, J.-K. Self-tuning experience-weighted attraction learning in games. J. Econ. Theory 2007, 133, 177–198. [Google Scholar] [CrossRef] [Green Version]
  8. Erev, I.; Roth, A. Predicting how people play games: Reinforcement learning in experimental games with unique mixed strategy equilibria. Am. Econ. Rev. 1998, 88, 848–881. [Google Scholar]
  9. Bala, V.; Goyal, S. A noncooperative model of network formation. Econometrica 2000, 68, 1181–1229. [Google Scholar] [CrossRef]
  10. Jackson, M.O.; Watts, A. The evolution of social and economic networks. J. Econ. Theory 2002, 106, 265–295. [Google Scholar] [CrossRef]
  11. Berninghaus, S.K.; Vogt, B. Network formation in symmetric 2 × 2 games. Homo Oecon. 2006, 23, 421–466. [Google Scholar]
  12. Berninghaus, S.K.; Ehrhart, K.-M.; Keser, C. Conventions and local interaction structures: Experimental evidence. Games Econ. Behav. 2002, 39, 177–205. [Google Scholar] [CrossRef]
  13. Brunner, C.; Camerer, C.F.; Goeree, J.K. Stationary concepts for experimental 2 × 2 games: Comment. Am. Econ. Rev. 2011, 101, 1029–1040. [Google Scholar] [CrossRef] [Green Version]
  14. Selten, R.; Chmura, T.; Goerg, S.J. Stationary concepts for experimental 2 × 2 games: Reply. Am. Econ. Rev. 2011, 101, 1041–1044. [Google Scholar] [CrossRef]
  15. Cassar, A. Coordination and cooperation in local, random and small world networks: Experimental evidence. Games Econ. Behav. 2007, 58, 209–230. [Google Scholar] [CrossRef]
  16. Kirchkamp, O.; Nagel, R. Naive learning and cooperation in network experiments. Games Econ. Behav. 2007, 58, 269–292. [Google Scholar] [CrossRef]
  17. Greiner, B. The online recruitment system ORSEE 2.0-A guide for the organization of experiments in economics. In Working Paper Series of Economics; University of Cologne: Cologne, Germany, 2004. [Google Scholar]
  18. Fischbacher, U. z-Tree: Zurich toolbox for ready-made economic experiments. Exp. Econ. 2007, 10, 171–178. [Google Scholar] [CrossRef]
  • 1Aspects of network formation, and learning in networks have been the center of attention in many studies (see, e.g., [9,10,11])
  • 2We thank one of our anonymous referees for suggesting this analysis.
  • 3Since we do not estimate parameters the fit of models using free parameters might improve if we would estimate parameters.
  • 4As benchmarks we use the predictions of the five equilibrium concepts reported by Selten and Chmura [1] and in Brunner et al. [13] and the predictions of the four learning models reported by Chmura et al. [2].
  • 5One can find detailed descriptions of the five concepts and the determination of the equilibria in Selten and Chmura [1] and in Brunner et al. [13].
  • 6As one can see, players at the edges of the lattice have their neighbors on the opposite edges.
  • 7We provided the same feedback as the players received in the experiment of Selten and Chmura [1].
  • 8All instructions were provided in German. An English translation of the written instructions is shown in the Appendix.
  • 9That the non-Nash concepts outperform Nash is in line with the results of SC and BCG.

Share and Cite

MDPI and ACS Style

Berninghaus, S.K.; Neumann, T.; Vogt, B. Learning in Networks—An Experimental Study Using Stationary Concepts. Games 2014, 5, 140-159. https://doi.org/10.3390/g5030140

AMA Style

Berninghaus SK, Neumann T, Vogt B. Learning in Networks—An Experimental Study Using Stationary Concepts. Games. 2014; 5(3):140-159. https://doi.org/10.3390/g5030140

Chicago/Turabian Style

Berninghaus, Siegfried K., Thomas Neumann, and Bodo Vogt. 2014. "Learning in Networks—An Experimental Study Using Stationary Concepts" Games 5, no. 3: 140-159. https://doi.org/10.3390/g5030140

Article Metrics

Back to TopTop