Repeated game
In game theory, a repeated game is an extensive form game that consists of a number of repetitions of some base game (called a stage game). The stage game is usually one of the well-studied 2-person games. Repeated games capture the idea that a player will have to take into account the impact of his or her current action on the future actions of other players; this impact is sometimes called his or her reputation. Single stage game or single shot game are names for non-repeated games.
Contents
Finitely vs infinitely repeated games[edit]
Repeated games may be broadly divided into two classes, depending on the horizon. The horizon is the player's belief about the number of repetitions of the stage game, and may be finite or infinite. The horizon is infinite in infinitely repeated games, so players in this type of game can expect to play indefinitely. By contrast, the horizon is finite in finitely repeated games, so players expect the game to terminate after playing the stage game some number of times. Note that for a game with an infinite horizon, the stage game will not necessarily be repeated an infinite number of times. A game repeated a finite number of times may be regarded as having an infinite horizon if the players in the game do not know how many times the game will be repeated. The results in these two cases are very dissimilar. The difference in strategies for finite versus infinite horizon games is a hotly debated topic, and many game theorists have differing views regarding it.
Infinitely repeated games[edit]
The most widely studied repeated games are games that are repeated an infinite number of times. In iterated prisoner's dilemma games, it is found that the preferred strategy is not to play a Nash strategy of the stage game, but to cooperate and play a socially optimum strategy. An essential part of strategies in infinitely repeated game is punishing players who deviate from this cooperative strategy. The punishment may be playing a strategy which leads to reduced payoff to both players for the rest of the game (called a trigger strategy). A player may normally choose to act selfishly to increase their own reward rather than play the socially optimum strategy. However, if it is known that the other player is following a trigger strategy, then the player expects to receive reduced payoffs in the future if they deviate at this stage. An effective trigger strategy ensures that cooperating has more utility to the player than acting selfishly now and facing the other player’s punishment in the future.
There are many results in theorems which deal with how to achieve and maintain a socially optimal equilibrium in repeated games. These results are collectively called "Folk Theorems". An important feature of a repeated game is the way in which a player's preferences may be modeled. There are many different ways in which a preference relation may be modeled in an infinitely repeated game, but two key ones are :
- Limit of means - If the game results in a path of outcomes and player i has the basic-game utility function , player i's utility is:
- Discounting - If player i's valuation of the game diminishes with time depending on a discount factor , then player i's utility is:
For sufficiently patient players (e.g. those with high enough values of ) , it can be proved that every strategy that has a payoff greater than the minmax payoff can be a Nash equilibrium - a very large set of strategies.
Finitely repeated games[edit]
Repeated games allow for the study of the interaction between immediate gains and long-term incentives. A finitely repeated game is a game in which the same one-shot stage game is played repeatedly over a number of discrete time periods, or rounds. Each time period is indexed by 0 < t ≤ T where T is the total number of periods. A player’s final payoff is the sum of their payoffs from each round.[1]
In each period of a finite game, players execute a certain amount of action. These actions lead to a stage-game payoff for the players. The stage game can be denoted by {A, u} where A = A1 * A2 *...* An is the set of profiles and ui(a) is player i’s stage-game payoff when profile a is played. The stage game is played in each period. Additionally, we assume that in each period t, the players have observed the history of play, or the sequence of action profiles, from the first period through period t-1. The payoff of the entire game is the sum of the stage-game payoffs in periods 1 through T. Sometimes, one should assume that all players discount the future, in which case we include a discount factor in the payoff specification.[2]
For those repeated games with a fixed and known number of time periods, if the stage game has a unique Nash equilibrium, then the repeated game has a unique subgame perfect Nash equilibrium strategy profile of playing the stage game equilibrium in each round. This can be deduced through backward induction. The unique stage game Nash equilibrium must be played in the last round regardless of what happened in earlier rounds. Knowing this, players have no incentive to deviate from the unique stage game Nash equilibrium in the second-to-last round, and so on this logic is applied back to the first round of the game.[3] This ‘unraveling’ of a game from its endpoint can be observed in the Chainstore paradox.
If the stage game has more than one Nash equilibrium, the repeated game may have multiple subgame perfect Nash equilibria. While a Nash equilibrium must be played in the last round, the presence of multiple equilibria introduces the possibility of reward and punishment strategies that can be used to support deviation from stage game Nash equilibria in earlier rounds.[3]
Finitely repeated games with an unknown or indeterminate number of time periods, on the other hand, are regarded as if they were an infinitely repeated game. It is not possible to apply backward induction to these games.
Examples of cooperation in finitely repeated games[edit]
X | Y | Z | |
A | 5 , 4 | 1 , 1 | 2 , 5 |
B | 1 , 1 | 3 , 2 | 1 , 1 |
Example 1: Two-Stage Repeated Game with Multiple Nash Equilibria
Example 1 shows a two-stage repeated game with multiple pure strategy Nash equilibria. Because these equilibria differ markedly in terms of payoffs for Player 2, Player 1 can propose a strategy over multiple stages of the game that incorporates the possibility for punishment or reward for Player 2. For example, Player 1 might propose that they play (A, X) in the first round. If Player 2 complies in round one, Player 1 will reward them by playing the equilibrium (A, Z) in round two, yielding a total payoff over two rounds of (7, 9).
If Player 2 deviates to (A, Z) in round one instead of playing the agreed-upon (A, X), Player 1 can threaten to punish them by playing the (B, Y) equilibrium in round two. This latter situation yields payoff (5, 7), leaving both players worse off.
In this way, the threat of punishment in a future round incentivizes a collaborative, non-equilibrium strategy in the first round. Because the final round of any finitely repeated game, by its very nature, removes the threat of future punishment, the optimal strategy in the last round will always be one of the game’s equilibria. It is the payoff differential between equilibria in the game represented in Example 1 that makes a punishment/reward strategy viable (for more on the influence of punishment and reward on game strategy, see 'Public Goods Game with Punishment and for Reward').
M | N | O | |
C | 5 , 4 | 1 , 1 | 0 , 5 |
D | 1 , 1 | 3 , 2 | 1 , 1 |
Example 2: Two-Stage Repeated Game with Unique Nash Equilibrium
Example 2 shows a two-stage repeated game with a unique Nash equilibrium. Because there is only one equilibrium here, there is no mechanism for either player to threaten punishment or promise reward in the game’s second round. As such, the only strategy that can be supported as a subgame perfect Nash equilibrium is that of playing the game’s unique Nash equilibrium strategy (D, N) every round. In this case, that means playing (D, N) each stage for two stages (n=2), but it would be true for any finite number of stages n.[4] To interpret: this result means that the very presence of a known, finite time horizon sabotages cooperation in every single round of the game. Cooperation in iterated games is only possible when the number of rounds is infinite or unknown.
Solving repeated games[edit]
In general, repeated games are easily solved using strategies provided by folk theorems. Complex repeated games can be solved using various techniques most of which rely heavily on linear algebra and the concepts expressed in fictitious play.
Incomplete information[edit]
Repeated games can include incomplete information. Repeated games with incomplete information were pioneered by Aumann and Maschler.[5] While it is easier to treat a situation where one player is informed and the other not, and when information received by each player is independent, it is possible to deal with zero-sum games with incomplete information on both sides and signals that are not independent.[6]
References[edit]
- ^ Knight, Vince. "Finitely Repeated Games". Game Theory. Retrieved 12/6/17. Check date values in:
|access-date=
(help) - ^ Waston, Joel (2013). Strategy: An Introduction to Game Theory. New York, London: W.W Norton and Company. p. 292. ISBN 978-0-393-91838-0.
- ^ a b Benoit, J.P. & Krishna, V. (1985). "Finitely Repeated Games". Econometrica: 905–922.CS1 maint: Multiple names: authors list (link)
- ^ Levin, Jonathan (May 2006). ""Repeated Games I: Perfect Monitoring"" (PDF). www.stanford.edu. Retrieved December 12, 2017.
- ^ Aumann, R. J.; Maschler, M. (1995). Repeated Games with Incomplete Information. Cambridge London: MIT Press.
- ^ Mertens, J.-F. (1987). "Repeated Games". Proceedings of the International Congress of Mathematicians, Berkeley 1986. Providence: American Mathematical Society. pp. 1528–1577. ISBN 0-8218-0110-4.
- Fudenberg, Drew; Tirole, Jean (1991). Game Theory. Cambridge: MIT Press. ISBN 0-262-06141-4.
- Mailath, G. & Samuelson, L. (2006). Repeated games and reputations: long-run relationships. New York: Oxford University Press. ISBN 0-19-530079-3.
- Osborne, Martin J.; Rubinstein, Ariel (1994). A Course in Game Theory. Cambridge: MIT Press. ISBN 0-262-15041-7.
- Sorin, Sylvain (2002). A First Course on Zero-Sum Repeated Games. Berlin: Springer. ISBN 3-540-43028-8.