| Literature DB >> 25873855 |
Abstract
The question of whether, and if so how, learning can be transfered from previously experienced games to novel games has recently attracted the attention of the experimental game theory literature. Existing research presumes that learning operates over actions, beliefs or decision rules. This study instead uses a connectionist approach that learns a direct mapping from game payoffs to a probability distribution over own actions. Learning is operationalized as a backpropagation rule that adjusts the weights of feedforward neural networks in the direction of increasing the probability of an agent playing a myopic best response to the last game played. One advantage of this approach is that it expands the scope of the model to any possible n × n normal-form game allowing for a comprehensive model of transfer of learning. Agents are exposed to games drawn from one of seven classes of games with significantly different strategic characteristics and then forced to play games from previously unseen classes. I find significant transfer of learning, i.e., behavior that is path-dependent, or conditional on the previously seen games. Cooperation is more pronounced in new games when agents are previously exposed to games where the incentive to cooperate is stronger than the incentive to compete, i.e., when individual incentives are aligned. Prior exposure to Prisoner's dilemma, zero-sum and discoordination games led to a significant decrease in realized payoffs for all the game classes under investigation. A distinction is made between superficial and deep transfer of learning both-the former is driven by superficial payoff similarities between games, the latter by differences in the incentive structures or strategic implications of the games. I examine whether agents learn to play the Nash equilibria of games, how they select amongst multiple equilibria, and whether they transfer Nash equilibrium behavior to unseen games. Sufficient exposure to a strategically heterogeneous set of games is found to be a necessary condition for deep learning (and transfer) across game classes. Paradoxically, superficial transfer of learning is shown to lead to better outcomes than deep transfer for a wide range of game classes. The simulation results corroborate important experimental findings with human subjects, and make several novel predictions that can be tested experimentally.Entities:
Keywords: agent-based modeling; connectionist modeling; cooperation and conflict; game theory; neural networks and behavior; transfer of learning
Year: 2015 PMID: 25873855 PMCID: PMC4379898 DOI: 10.3389/fnins.2015.00102
Source DB: PubMed Journal: Front Neurosci ISSN: 1662-453X Impact factor: 4.677
Game class characteristics.
| Zero-sum | ZS | 0 or 1 | 1 or 0 | Possibly | Pure conflict | |
| Prisoner's Dilemma | PD | 1 | 0 | Yes | Social dilemma | |
| Mixed strategy | MS | 0 | 1 | No | Discoordination | |
| Stag hunt | SH | 2 | [1, 1] | 1 | No | Coordination |
| Chicken | CH | 2 | [0, 1] | 1 | No | Anti-coordination |
| Battle of the Sexes | BOS | 2 | [0, 1] | 1 | No | Coordination |
| No competition | NC | 1 | 0 | Possibly | No conflict | |
PSNE, pure strategy NE; MSNE, mixed-strategy NE; PDNE, payoff-dominant NE; RDNE, risk-dominant NE.
Figure 1Convergence of the simulations—mean payoffs as a function of the number of games played.
Agent heterogeneity—rank correlation of choice probabilities and % choice agreement.
| ZS | 0.96 | 0.94 | 0.98 | 80 | 77 | 82 |
| PD | 0.83 | 0.67 | 0.93 | 100 | 99 | 100 |
| MS | 0.01 | −0.54 | 0.52 | 50 | 45 | 59 |
| SH | 0.87 | 0.79 | 0.94 | 100 | 99 | 100 |
| CH | 0.55 | 0.19 | 0.90 | 61 | 57 | 65 |
| BOS | 0.86 | 0.69 | 0.96 | 87 | 85 | 90 |
| NC | 0.90 | 0.84 | 0.94 | 100 | 99 | 100 |
| ALL | 0.93 | 0.89 | 0.95 | 72 | 68 | 76 |
Spearman rank correlation of agents' behavior by training sets.
The higher a cell's value, the darker the shading.
Mean payoffs (conditional on the training and test sets).
The higher a cell's value, the darker the shading.
Figure 2Mean payoffs (conditional on the training and test sets).
Joint probability of playing a PSNE (conditional on the training and test sets)—Games with a unique PSNE.
The higher a cell's value, the darker the shading.
Joint probability of playing a PSNE (conditional on the training and test sets)—Games with two PSNE.
The higher a cell's value, the darker the shading.
Figure 3Probability of playing Nash equilibria.
Probability of risk- vs. payoff-dominant equilibria in SH games.
The higher a cell's value, the darker the shading.
Transfer of learning in games with a unique PSNE (probability of joint PSNE play).
| Training | PD | 0.998 | 0.004 | 0.532 |
| NC | 0.008 | 0.999 | 0.413 | |
| ALL | 0.901 | 0.844 | 0.83 |
A sequence of games spanning four game classes.
Shaded cells correspond to pure-strategy Nash equilibria of the games λ = {0.01, 0.02, …, 0.23, 0.24}.
The higher a cell's value, the darker the shading.
Figure 4Deep vs. superficial transfer of learning in a sequence of games.