Block 7
Block 7
~
ignou
THE PEOPLE'S
UNIVERSITY
MEC -101
MICROECONOMIC
Indira Gandhi National Open University
School of Social Sciences ANALYSIS
Volume 2
--~~
MEC-IOl
1te:\11 ~~ ..~
~ 'UNIVERS1TY
Microeconomic
Indira Gandhi Analysis
National Open University
School of Social Sciences
Block
7
NON-COOPERATIVE GAME THEORY-I
UNIT 20
Static Games of Complete Information 5
UNIT 21
Static Games with Complete Information: Applications 45
UNIT 22
Dynamic Games with Complete Information 77
Expert Committee
Prof. Bhaswar Moitra
Prof. Gopinath Pradhan
Department of Economics
School of Social Sciences
Central Road, Jadavpur University
Indira Gandhi National Open University
Kolkata
New Delhi
Dr. Naresh Kumar Sharma
Prof. Narayan Prasad
University of Hyderabad
School of Social Sciences
School of Economics, Hyderabad
Indira Gandhi National Open University
Dr. Anirban Kar New Delhi
Deptt. of Economics
Prof. Kaustuva Barik
Delhi School of Economics
School of Social Sciences
University of Delhi
Indira Gandhi National Open University
Dr. Indrani Roy Chowdhury New Delhi
Economics Faculty
Prof. B.S. Prakash
Jamia Millia Islamia, New Delhi
School of Social Sciences
Prof. Sudhir Shah Indira Gandhi National Open University
Deptt. of Economics, Delhi School of Economics New Delhi
University of Delhi
Mr. Saugato Sen
Dr. Manish Gupta School of Social Sciences
NIPFP, New Delhi Indira Gandhi National Open University
New Delhi
PRINT PRODUCTION
Sh. Tilak Raj
Assistant Registrar
MPDD, IGNOU, New Delhi
September 2021 (Reprint)
© Indira Gandhi National Open University 1993
ISBN: 978-93-86375-97-1
All rights reserved. No part of this work may be reproduced in any form, by mimeograph or any other
means, without permission in writing from the lndira Gandhi National Open University.
Further information about the lndira Gandhi National Open University courses may be
obtained from the University's office at Maidan Garhi, New Delhi-ll0068
Printed and Published on behalf of the Indira Gandhi National 'Open University, by Registrar
MPPDD, IGNOU New Delhi.
Printed at : Mls Nova Publications and Printers (Pvt.) Ltd., Plot No. 09-10, Sector-59, Phase-2 Faridabad-121004,
Phone: 0129-4317645.
BLOCK 7 NON-COOPERATIVE GAME
THEORY-I
I. Introduction
This block deals with the non-cooperative game theoretic approach often used in
seeking solutions to economic problems in general and microeconomic problems in
particular. Unit 20 introduces the static games with complete information along with
the basic concepts that helps understand the method of solution. The concept of
Nash equilibrium is introduced by considering pure and mixed strategies. The unit
also contains the description of the concept of equilibrium and representation of
games through normal form. Some of the applications of static games with complete
information are discussed in Unit 21. Insights thrown by game theory to problems
of tragedy of commons and duopoly markets are important themes covered in the
unit. The last unit of the block (Unit 22) contains solution to problems when we
have situation of dynamic games with complete information. Itintroduces the concept
of backward induction as a refinement ofNash equilibrium and presents the extensive
format of a game.
UNIT 20 STATIC GAMES OF
COMPLETE INFORMATION
Structure
20.0 Objectives
20.1 Introduction
20.2 Some Examples of Strategic Games
20.3 Classifications of Games
20.3.1 Cooperative and Non-Cooperative Game Theory
20.3.2 Games of Complete Information and Incomplete Information
20.3.3 Strategic (or Normal Farm) and Extensive Form Games
20.4 Rules of the Game
20.5 Normal Form of Game under Complete Information
20.5.1 Some Well-known Strategic Games
20.5.1.1 Matching of Pennies
20.5.1.2 Battle of Sexes
20.5.2.3 Prisoners' Dilemma Game
20.5.1.4 Hawk-Dove (or Chicken) Game
20.5.1.5 Coordination Game
20.0 OBJECTIVES
After going through this unit, you will be able to understand:
• various classifications of games like cooperative and non-cooperative
games along with that of simultaneous move games;
5
Non-Cooperative • rules of strategic game or normal form of game;
Game Theory-I
• basics of the normal form or the strategic form of game;
• solution concept under Dominant Strategy;
• solution concept under Nash equilibrium in pure strategy; and
• solution concept under Mixed Strategy Nash equilibrium.
20.1 INTRODUCTION
Game theory studies the general principles that explain how people and
organisations act in strategic situations in a multiple agent framework.
A strategic situation involves when two or more interacting players, who
make decisions by trying to anticipate the actions and reactions of others.
Thus, Game Theory can be described as a strategy theory, or theory of
interactive decision making. It is rather a formal study of interdependence,
conflict and cooperation. Game theory can be used to explain phenomena in
the areas of economics, political theory, evolutionary science and military
strategy, to name a few.
Game theory was established as a field in its own right after the 1944
publication of the monumental volume 'Theory of Games and Economic
Behavior' by von Neumann and the economist Oskar Morgenstem. It
provided much of the basic terminology and problem setup that is still in use
today. In 1950, John Nash demonstrated that [mite games have always have
an equilibrium point, at which all players choose actions which are best for
them, given their opponents' choices. This central concept of non-cooperative
game theory has been a focal point of analysis since then. Since the 1970s, it
has driven a revolution in .economic theory. Additionally, it has found
applications in sociology and psychology and established links with evolution
and biology.Game theory received special attention in 1994 with the awarding
of the Nobel Prize in economics to Nash, John Harsanyi and Reinhard Selten.
Its formalism comprises a set of agents and a function that returns the
value each subset of agent (i.e., each coalition) can create on its own (the
characteristic function). The characteristic function is the input to a
solution concept, which returns the value captured by each agent (their
imputation). Because no structure is put on the bargaining procedure, the
solution concept models the outcome of the negotiations occurring
among the agents, accounting for the value each coalition can create.
While the solution concepts are the Core and the Shapley value, but a
number of others have been developed (for further details see Osborne
and Rubinstein (1994) and Owen (1995)).
The core is the set of imputations such that each coalition receives at
least as much as it can create on its own. This concept formalizes a
notion of individual and coalitional self-interest. The core is thus
appropriate to model the competitive nature of free-form competitive
interactions. However, the core may not exist and is not necessarily
unique. Non-existence of the core signals the intrinsic instability of a
particular interaction. Non-uniqueness of the core implies that
competitive forces alone are not enough to determine each agent's value
capture and that negotiating abilities also come into play. An agent's
added value (or marginal contribution) is the value lost to the grand
coalition (the coalition of all agents) if the agent is not included. The
added-value principle, implied by the core, states that under competitive
free- form interaction an agent can capture no more than her added value.
Otherwise, the other agents would be better off to exclude her and trade
only among themselves.
The Shapley value is the average added value of an agent to all possible
coalitions. It always exists and is unique. It emphasizes the fairness of
the division of value among agents. It has been used to model the
allocation of costs among users of a common facility and to measure
political power.
8
Static Games of
20.3.2 Games of Complete Information and Incomplete
Complete
Information Information
Non-Cooperative games can be classified into (i) Games of Complete
Information and (ii) Games of Incomplete information.
Under non-cooperative games, a category of game in which knowledge about
other players, utility functions, payoffs, strategies and "types" of players are
common knowledge i.e., those information are available to all participants.
Then such games are called Games of Complete Information.
The extensive form (also called the sequential form), also called a game tree,
is more detailed than the strategic form of a game. It is a complete description
of how the game is played over time. This includes the order in which players
take actions, the information that players have at the time they must take those
actions and the times at which any uncertainty in the situation is resolved. A
game in extensive form may be analysed directly, or can be converted into an
equivalent strategic form.
9
Non-Cooperative Check Your Progress 1
Game Theory-I
1) What is a cooperative game?
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.......•••••••....................• ! •••••••.......•••••••••••.....•••••.....••••.•.....••
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
.........................................................................................
10
6) What is a non-cooperative game? Static Games of
Complete
Information
.................. ' .
..........................................................................................
10)There are two players, 1 (kicker) and 2 (goalie). Each has two actions,
ai E {L, R} to denote left or right. The kicker wins when they choose
opposite directions while the goalie wins if they choose the same
direction. Using 1to denote a win and -1 to denote a loss, write the game
in a matrix form.
..........................................................................................
11
Non-Cooperative 11) Provide an example of a game in which there is more than one weakly
Game Theory-I dominant strategy equilibrium.
............... ; .
13) Provide an example of a game in which there is more than one weakly
dominant strategy equilibrium.
5) How much are the players in a game supposed to know about the rules?
In game theory it is standard to assume common knowledge about the
rules. Everybody has a knowledge about the rules means that if you asked
any two players in the game, a question about who, what, when or how
much, they would give you the same answer.
The two principal representations of (the rules of) a game are called,
respectively, the Normal (or strategic) form of a game and the Extensive form
ofa game.
12
Static Games of
20.5 NORMAL FORM OF GAME UNDER Complete
COMPLETE INFORMATION Information
Suppose the game 'consisting of n players and let the representative player is
denoted by player I ( i = 1, 2, n). Let Si denotes the set of strategies
available to each player i and let si denotes the an arbitrary specific strategy
(Si E Sa, to indicate the strategy Si is the member of the set of strategy Si' Let
(Sv S2, S3' ... .;... sn)denotes a combination of strategies one for each player, let
Ui denote the player i's payoff Ui(SVS2, sn), where players choses the
strategies (sv S2' sn). Thus the normal form of n-players game is
specified as G = {Sv S2' Si, Uv U2, Ui} (20.1).
13
r
Two members «1 and 2) of a criminal gang are arrested and imprisoned. Each
prisoner is in solitary confinement with no means of communicating with the
other. Prosecutors lack sufficient evidence to convict the pair on the principal
charge. They are being interrogated separately about their crime. Each
prisoner is given the opportunity either to: betray the other by testifying that
the crime is committed, or to cooperate with the other by remaining silent.
The convicts have the options either to confess or not to confess their crime.
Based on their joint decision penalty will be decided. If both of them confess
the crime will be established and each will get 6 years of imprisonment. If
none of them confesses, the crime is not established. So, they will be
nominally punished for 1 year. However in case one confesses and the other
does not, the crime is established, so the convict who has confessed will be set
free for her honesty and the other one will be put under the bar for 9 years.
Table 20.3.
14
Table 20.4: Prisoner's Dilemma in Generalized Form Static Games of
Complete
Player 2 Information
The condition that must hold for a prisoner's dilemma game, in the strong
sense, is that.if the payoffs follows the inequality: ·T > R > P > S. The payoff
relationship R > P implies that mutual cooperation is superior to mutual
defection, while the payoff relationships T > Rand P > S imply that defection
is the dominant strategy for both agents. That is, mutual defection is the
unique Nash equilibrium in the game (i.e., the only outcome from which each
player could only do worse by unilaterally changing strategy). The dilemma
then is that mutual cooperation yields a better outcome than mutual defection
but it is not the rational outcome because from a self-interested perspective,
,"
the choice to cooperate, at the individual level, is irrational. This sort of
outcome is described by Mancur Olson (1965) as - individual rationality
produces collective irrationality.
The prisoner's dilemma game is widely used in the study of game theory.
The name "chicken" has its origins in a game in which two drivers drive
towards each other on a collision course: one must swerve, or both may die in
the crash, but if one driver swerves and the other does not, the one who
swerved will be called a "chicken", meaning a coward; this terminology is
most prevalent in political science and economics.
Player 2
Swerve (Sw) Straight (St)
Player 1 Swerve (Sw) 0,0 -I, +I(NE)
Two versions of the payoff matrix for the Chicken game are presented here
(Tab1e-20.5 and Table 20.6). Two (young) players are engaged in a conflict
situation. They may be racing their car towards each other on Main Street,
while being egged on by their many friends. If player 1 hangs tough by racing
straight and stays in the Centre of the road while the other player swerves by
moving out of the way, then all glory is her and the other player eats humble
pie (St, Sw). If they both hang tough by racing straight (St, St) they end up
with broken bones, while if they both swerve they have their bodies- but not
their pride - intact.
In such game the outcomes are such that, where each player would prefer to
win over tying, prefer to tie over losing and prefer to lose over crashing. Here,
the benefit of winning is 1, the cost of losing is -I and the cost of crashing is -
10. It is mutually beneficial for the players to play different strategies.
Hawk-Dove Game
In the biological literature, this game is referred to as Hawk-Dove. The earliest
presentation of a form of the Hawk-Dove game was by John Maynard Smith
and George Price in their paper, ''The logic of animal conflict". Both in
Chicken and Hawk-Dove, it is mutually beneficial for the players to play
different strategies. It can be thought of as the opposite of a coordination game
(see below for Coordination game), where playing the same strategy Pareto
dominates playing different strategies. The underlying concept is that players
use a shared resource. In coordination games" sharing the resource creates a
.. benefit for all: the resource is non-rivalrous and the shared usage creates
positive externalities. In anti-coordination games the resource is rivalrous but
non-excludable and sharing comes at a cost (or negative externality).
The traditional payoff matrix for the Hawk-Dove game is given in Table 20.7,
where V is the value of the contested resource and C is the cost of an escalated
fight. It is (almost always) assumed that the value of the resource is less than
the cost of a fight, i.e., C > V > 0. If C :s V, the resulting game is not a game'
of Chicken but is instead a Prisoner's Dilemma. The exact value of the Dove
vs. Dove payoff varies between model formulations. Sometimes the players
are assumed to split the payoff equally (V 12 each), other times the payoff is
assumed to be zero. .
16
While the Hawk-Dove game is typically taught and discussed with the payoffs Static Games of
in terms of V and C, the solutions hold true for any matrix with the payoffs in Complete
Information
Table 20.8, where W > T > L > X.
Table 20.7: Hawk-Dove game
Player 2
Hawk (H) Dove (D)
Hawk (H) (V-C)/2, (V-C)/2 V,O
Player 1
Dove (D) 0, V V/2, V/2
17
Non-Cooperative 20.6.1 Dominance
Game Theory-I
Consider a player have two strategies A and B so that, given any combination
of strategies of the other players, the outcome resulting from A is better than
the outcome resulting from B. Then strategy A is said to dominate strategy B.
Since all players are assumed to be rational, they make choices which result in
the outcome they prefer most, given what their opponents do. If out of N
strategies A, B, C, ... , N, the outcome from strategy A is better than any other
(N-I) strategies, irrespective of the strategies of the other players then there is
a compelling reason to think that the rational player will choose it. If a
strategy is bad for an agent no matter what the others might be doing there is
equally compelling reason to think that a rational player would avoid it. A
rational player will never choose to play a dominated strategy.
In this above example of strategic game the Player's 1 strategy Low (L) is
dominated by High (H) if:
a) 1Ti (H, N) ~ 1Ti (L, N) (20.3a)
b) 1Ti (H, S) ~ 1Ti (L, S), (20.3b)
The first inequality says that H yields a higher payoff than L to the player 1, if
the player 2 plays her first strategy i.e., North (N). The second inequality says
the same is true even if player 2 plays her second strategy.
With at least one of those inequalities being strict, Low is not dominated by
High (and vice versa).If it is better against, say, South but does worse against
North i.e.,
a) 1Ti (H, N) > 1Ti (L, N), (20.4a)
18 But b) 1Ti (H, S) < 1Ti (L, S), (20.4b).
Consider the Prisoners Dilemma game of Table 20.3. Static Games of
Complete
1) Each agent has 2 strategies to cooperate (not confess) with one's rival and Information
the other to defect from cooperation (by confessing).
2) Each player has a dominant strategy (to confess) or to defect from the
cooperation strategy.
3) Dominant strategy equilibrium is worse for both the agent than the non
equilibrium situation or the dominated strategy to cooperate with the rival
(not confess, not confess).
The strategy confess is dominating the strategy defect, as it gives the Convict
.1 a higher payoff than Defect irrespective of the strategy played by the
Convict 2 i.e., whether she confesses or defects ~
(07rI(C,C) = (-6) > 7r1(d,c) = (-9)
(ii)7r1(c, d) = (0) > 7r1(d, d) = (-1)
Similarly, for the Convict 2 also, confession is dominating the strategy defect,
as it gives the Convict 2 a higher payoff than Defect, irrespective of the
strategy played by the Convict 1 whether she confesses or defects ~.
(iii)7r2(C, c) = (-6) > 7r2(d, c) = (-9)
(iv) 1r2(C, d) = (0) > 7r2(d, d) = (-1).
So, in this Prisoner's Dilemma game, confession is the Dominant strategy for
both the convict. Thus in this game, it is reasonable to predict that both the
players will end up confessing in the equilibrium. Hence {Confess, Confess}
. is the Dominant Strategy Equilibrium.
Player 2
Left Right
Player 1 Up (1,1) (0,1)
Middle (0,2) (1,0)
Down (0,-1) (0,0)
For the Player 1, the strategies Up and Middle dominate the strategy Down.
Hence Down is the dominated strategy for the player 1. So, for a rational
player, it will be irrational to play a dominated strategy Down. The reason is
that by playing any strategy that dominates (this dominated strategy), the
player can guarantee a payoff which is at least as high, irrespective of the
moves of the rival players. Hence the player 1 will ever play Down but will
rather play either Up or Middle.
The logic could then 'Set in motion a chain reaction. In a game once it is
known that the player 1will not play the dominated strategy Down, other
player might find that that certain of their strategies are in fact dominated.
19
Non-Cooperative This is because Player 2 for instance, no longer has to worry how her
Game Theory-I strategies would perform against player l' s dominated strategies. So, some of
the strategies of player 2's, which are only good against player 1's dominated
strategies, might turn out to be bad strategies. In that case player 2 will not
play those strategies. This might lead to a third round of discovery of bad
strategy by some of the other players and so on.
Table 20.11a
Player 2
Player 1 Left Right
Up (1,1) (0,1)
Middle (0,2) (1,0)
.Thus if it is known to Player 2 that Player 1 will never play Down, the Right
looks dominated to her. Therefore the rational Player 2 would never play
Right. But then the Player 1 should not worry about Player 2 playing Right.
. Hence she should choose the strategy Up.
Table 20.11b
Player 2
Left
Player 1 Up (1,1)
Middle (0,2)
Table 20.11c
Player 2
Left
Player 1 Up (1,1)
Player 2
Left (L) Centre (C) Right (R)
Player 1 Top (T) (0,4) (4,0) (5,3)
Middle (M) (4,0) (0,4) (5,3)
Bottom (B) (3,5) (3,5) (6,6)
20
In this above game (Table20.12) there are no strictly dominated strategies to Static Games of
be eliminated. Since all the strategies in the game survive IEDS, the process Complete
Information
produces no prediction whatsoever about the play of the game.
................ , .
Player 2
left middle right
Player 1 up 1,0 1,2 0,1
down 0,3 0,1 2,0
The above equation (20.3) says that each player i, in playing si .is playing a
best response to the other players' strategy choice. This condition includes
two requirements of Nash equilibrium:
1) Each player must be playing a best response against a conjecture
2) The conjecture must be correct.
Thus Nash equilibrium includes the first requirement because si is the ith
player's best response against the conjecture s~i(optimal strategy choice of all
the other (n-l) players (except the ith player». It includes the second
requirement because no player has an incentive to change her strategy (from
si). Hence si is stable and each player's conjecture is correct.
Suppose game theory offers the strategies (s~, s;, ..... , s~) as a solution to the
normal form of game G = {Sv S2, ..... Si, Uv U2, ..... ud. Saying that
(s~, s;, ..... , s~) is not a Nash equilibrium of G is equivalent of saying that
there exists some player i such that s; is not a best response to
(s~, s;, ... S;-1' S;+1'" s~). That is there exists some s;'inSi, such that
Thus if the theory offers the strategy (s~, s;, ..... , s~), as solution but these
strategies are not a Nash equilibrium, then atleast one player will have an
incentive to deviate from the theory's prediction, so the theory will be
falsified by the actual play of the game.
Consider the case of two player 1 and.2, each with two strategies (av a2) for
the player 1 and (bv b2) for the player 2. For example, (az, bt) is a Nash
equilibrium if and only if :
(a). 1f1 (a2, bt) ~ 1f1 (av bt) (20.8a)
(b).1f2(az,bt) ~ 1fz(az,bz), (20.8b)
In contrast, situations where participants can all gain or suffer loss together
are referred to as non-zero-sum. Thus in non-zero-sum games are games in
which the sum of gains and losses by the players are sometimes more or less
than what they began with.
In the equilibrium of the zero sum game, each player is choosing her or her
best response, given her beliefs about what the others will do. In anticipating
such best responses, each player will expect to receive the worst payoff
associated with each action and will choose her own action accordingly. She
is thus choosing her best payoff from among the set of worst pay offs.
In some game most notably in sports contexts, the interests of the two sides
are exactly opposite to each other. Then, for each combination of players'
choices, the payoff of one can be obtained by reversing the sign of the payoff
of the other. So, the payoff matrix under a zero sum game can be
simplistically represented by showing the payoff of one player. Consider the
following example of zero sum game where in a soccer game the two players
(Offense and Defense) are represented row wise and column wise with the
strategies respectively (Run, Short Pass, Medium Pass and Long Pass) and
(Run, Pass, Blitz). Suppose the payoff table shows the row player's (Offense)
payoff and therefore it is implicit that the Column player's (Defense) payoff
will be the negative of the same. The Offense wants the outcome to be a cell
with as high a number as possible in each column. The Defense (column)
wants the outcome of the cell with as Iowa number as possible. See Table 20-
20.13.
Using this logic the Offense figures that for each of her row, the defense or the
Column will choose the column with the lowest number in that row. Therefore
the Row should choose the row that gives her the highest among the lowest
numbers i.e., maximum among the minima - the maxi-min strategy.
Similarly, the Columns reckon that, for each of her columns, Row will choose
the row with the largest number in that column. The Column should choose
the column with the smallest number among the largest one or the minimum
among the maxima - called the mini-max strategy. If the Offense (row's)
maximin values and the defense's minimax values are in the same cell of the
payoff matrix, then the outcome is the Nash equilibrium of the zero sum
game - a concept which we will elaborate in details in the next section. The
method of finding equilibrium in the zero sum game is called maximin-
minima x method. The equilibrium will lead to Nash equilibrium in pure
strategies, if it exists.
We begin by finding the offense's worst payoff from each her strategy (which
is lowest number in each row) and defense's worst payoff in each strategy
(which should be the highest number from each column). Thus the offense's
worst payoff from Run is 2, worst payoff from Short Pass is 5.6, worst payoff
from Medium Pass is 1 and worst payoff from Long Pass is -2. We write the
minimum of each row at the last column. The Defense's worst payoff from
Run is 10, worst payoff from Pass is 5.6 and worst payoff from Blitz is 13.
We write the maximum of each column at the bottom of the column.
23
Non-Cooperative The next step is to find out the best of each player's worst possible outcomes,
Game Theory-I i.e., the largest row minimum and the smallest column maximum. The largest
row minima is 5.6, so the offense can ensure itself a gain of 5.6 yards by
playing the Short Pass, this is her maximin. The lowest of the column maxima
is 5.6, so the defense can be sure of holding the offense down to a gain of 5.6
yards by deploying its Pass Defense. This is the defense' s minimax.
Now from this two given strategy choices, we see the maxim in and the
minimax values are belonging to the same cell of the game table. Thus offense
maximinis the best response to the defense' s minimax strategy and vice versa,
which is basically the Nash equilibrium of the game. The equilibrium entails
the offense attempting a Short Pass while the defense defends against a Pass.
A total 5.6 yards will be gained by the offense (and given up by the defense).
If the minimax method fails to fmd an equilibrium in some zero sum game,
then we conclude that the game has no Nash equilibrium is pure strategies (we
will address the concept of mixed strategy equilibrium in the next section).
Moreover this minimax method cannot be applied to the non zero sum games.
As in those games the opponent's best is not necessarily the defense's worst.
a) Play Prescription
One can think of a Nash equilibriums*as a prescription for play. If the
strategy vector is proposed to the players, then it is stable prescription in
the sense that no one has an incentive to play otherwise. By playing an
alternative strategy, a player would simply lower her payoffs, if she
thinks that others are going to follow their part of the prescription.
b) Pre-play Communication
How would the players in a game find their way to a Nash equilibrium?
One answer that has been proposed is that they could coordinate on a
Nash equilibrium by way of pre-play communication,· i.e., they could
coordinate by meeting before (the game is actually played) and discussed
24 their options. It is not credible for the players to agree on anything that is
not a Nash equilibrium because atleast one player would be worse off Static Games of
against such agreement. Complete
Information
c) Rational Introspection
A related motivation is rational introspection: each player could ask
himself what she expects will be the outcome to a game. Some candidate
outcome will appear unreasonable in that there are players who could do
better than they are doing, i.e., there will be players not playing a best
response. The, only time no players appears to be making mistake is when
each is playing a best response, i.e., when we are at a Nash equilibrium.
d) Focal Point
Another motivation is the idea that a Nash equilibrium forms a focal
point for the players in a game, The intuitive idea of a focal point was
first advanced by Thomas Schelling in 1960 in his book: The Strategy of
Conflict. It refers to a strategy vector that stands out from the other
strategy vectors because of some distinguishing characteristics. Consider
the two Coordination game: (i) two players would have to write down
either heads or tails. They will be paid if their choices match. (ii) To
players have to meet in Canaught Place, New Delhi and have to choose a
time for their meeting. Again they are paid only when their chosen time
coincide. Schelling found that in a disproportionate number of cases,
students choose head in the case (i) and 12:00 pm in the case of (ii). A
. Nash equilibrium strategy vector is a focal point because it has the
distinguishing characteristics that each player plays a best response under
that strategy vector.
The best response to the player 1 i.e., Husband, is to play (F) against his
wife's choice of (F) and earns a payoff of 3. If the Husband deviates and
play (M) against his Wife choice of (F) he is worse off as he gets 0
(because 3 > 0). Similarly, playing F is the optimal strategy for Wife
(with a payoff 1) against her Husband's strategy of (F). If she deviates
and play (M) against her husband's choice of (F), she is worse off as
compared to the strategy choices (F, F), because 1>0.
Similarly, the best response to the Husband, is to play (M) against his
wife's choice of (M) and earns a payoff of 1. If the Husband deviates and
25 '
Non-Cooperative play (F) against his Wife choice of (M) he is worse off as he gets 0
Game Theory-I (because 1 > 0). Similarly, playing M is the optimal strategy for Wife
(with a payoff 3) against her Husband's strategy of (M). If she deviates
and play (F) against her husband's choice of (M), she is worse off as
compared to the strategy choices (M, M), because 3>0.
So, (H, H) and (M, M) are the two best response strategies of the couple
and are therefore Nash equilibrium.
If both the convicts play Non-Confession (NC, NC) the suspected crime
cannot be established' because of the absence of eye witness and therefore
they would get a nominal punishment of 1 year imprisonment
(rrl(NC,NC) = O,rrz(NC,NC) = 0). But if any of the players defects
from (NC, NC) and confesses his payoff would increased as he would be
set free for his honesty (asrrl (C, NC) = 0 andrrz(NC, C) = 0). So for
each of the player, there is an incentive to deviate from the strategy Not
Confess (NC). Moreover, if any of the player ditches from NC (given the
other stick to NC) it becomes really costly for the player who has not
deviated. Thus if the Convict 1 deviates from the (NC) and plays C,
whereas the Convict 2 plays NC, it is Convict 2 who will be' punished for
9 years of imprisonment and the Convict 1 will be set free (i.e.,
rrz (C, NC) = -9, rrl (NC, NC) = 0,). Similarly, you may check the case
when the Convict 2 deviates from (NC) and plays Confess (C), where as
the Convict 1 plays (NC). In that case the Convicts 1's payoff will be 9
years of imprisonment and the Convict 2 will be set free (rrl (C, NC) =
-9), rrz(NC, C) = 0)).
For any matrix with the payoffs in Figure 20.8, where W > T > L > X, the
Hog Dove game has multiple Nash equilibrium i.e., (H, D) and (D, H).
Consider the strategy (H, D). Given that Player I plays H, it is optimal
for the Player 2 to play D because his payoff from playing D (i.e., L)
exceeds his payoff from playing H (L > X). Similarly, given Player 2
plays D, it is optimal for the Player 1 to play H because his payoff from
playing H (i.e., W) exceeds his payoff from playing H (which is X) since
(W > X). Therefore the strategy (H, D) is mutually the best response
strategy or the Nash equilibrium. You may also check the strategy (D, H),
by the similar reasoning it is the mutually best response strategy or the
N ash equilibrium.
You may also check why the strategies like (H,H) or (D, D) are not the
Nash equilibrium. In such cases, given the other player's strategy, each of
the players has an incentive to deviate.
4) Coordination Game
In game theory, coordination games are a class of games with multiple
. pure strategy Nash equilibria in which players choose the same or
corresponding strategies. Consider the ,Minimum-Effort Coordination
game given in Figure 20.9. In this game it is optimal for both the players'
to play the same strategy i.e., putting low effort (1,1) or high effort (2, 2)
as the best response strategy at the equilibrium. So, (1,1) or (2, 2) are the
Nash equilibrium strategy.
Notice that the low effort outcome (1, 1) is an equilibrium, since a costly
increase in effort by only one person will not raise the amount produced.
Given Player 1 plays to put low effort(l), it is optimal for the player 2, to
put low effort, because (1- c) > (1- 2c), with 0 < c <1. Similarly, given
Player 2 plays to put low effort, it is optimal for the player 1 to put low
effort because (1- c) > (1- 2c). As long as c is less than 1, payoffs are
maximized when all players choose the highest possible effort. So, high
effort outcome (2, 2) is also mutually best response strategy and therefore
a Nash equilibrium. In this case you may check that a reduction in effort
by only one person will lower the minimum by more than the cost
savings c.
Note that any common effort level constitutes a Nash equilibrium, since a
costly unilateral increase in effort will not raise the minimum and a
unilateral decrease will reduce the minimum by more than the cost when
c < 1. However, this argument does not depend on the number of players,
so noncritical changes in c and n will not alter the set of Nash equilibria,
despite the reasonable expectation that efforts should be high for
sufficiently low effort costs and low numbers of participants. See Table-
20.9.
, Consider a 2 player 3 strategy game. Suppose the strategy (a2, b3) is the IEDS
solution. In order to establish that this pair is also a Nash equilibrium, we need
to show that a2 has a higher payoff than a1 or a3 when played against b3,
Likewise b3 has a higher payoff than b2 or b1 against player l' s strategy a2,
Suppose a1 is the strategy of the player 1 that is eliminated in the last round of
elimination and likewise b1 is the last round of elimination. We can now
proceed by the following three step proof:
Step 1: Regardless of whether player 2 plays b30r b'', a2 has a payoff as high
as ai, In particular against b 3, a 2 is at least as good a response as ai,
How is a2 's payoff compare with those of a3? Since 3 got eliminated in the
very first round it must have been dominated by either/ 1. If it is dominated by
2 we are done. (Why?) If it is dominated by 1, then we know that it does
worse than 1, against player 2's strategy of 3, But in turn 1 has no higher a
payoff than 2 against that same strategy 3, We have therefore concluded:
Step 3: For player 12 is the best response against+, Likewise+ is the best
response against 2,
Table 20.14
Player 2
1 2 3
Player 1 1
2
3
The reverse implication is not necessarily true; there may be N ash equilibria
that are not the solution to IEDS, This may be easy to see in any game, such
as Battle of Sexes, in which there is a Nash equilibrium but the game is not
dominance solvable. Another situation where Nash does not imply IEDS can
arise when there are multiple Nash equilibria; some of these Nash equilibria
would be eliminated by iterated elimination of dominant strategies.
28
Static Games of
Check Your' Progress 3 Complete
Information
1) What is minimax strategy?
.......................................................................................
......................................................................................
......................................................................................
......................................................................................
......................................................................................
L R
T (3,1) (2,-20)
M (5,0) (-10,1)
B (-10,0) (24,2)
......................................................................................
......................................................................................
......................................................................................
. .
......................................................................................
3) How do you fmd a relation between elimination of dominated strategies
(IEDS) and Nash equilibrium .
......................................................................................
......................................................................................
......................................................................................
......................................................................................
i) (F, F)
ii) (M, M) and
iii) A coin toss i.e., a strategy with equal likelihood to going to Football or
Movie. .
29
r
Non-Cooperative Note that the coin toss has clearly enlarged the set of initial choices as strategy
Game Theory-I (iii). is evidently distinct from (i) and (ii). Strategies (i) and (ii) as well as
every other strategy discussed so far are called Pure Strategies. A strategy like
(iii) is called the Mixed Strategy. In fact strategy (iii) is just one of the many
mixed strategies.
The obvious question now is how a mixed strategy payoff will be evaluated?
It follows that the right way to evaluate the uncertainty inherent in .a mixed
strategy is to take its expected payoff. Expected payoffs are computed in two
steps:
Step 1: Weight the payoff to each pure strategy by the probability with which
that strategy is played.
Step 2: Add up the weighted payoffs.
Let us recall the Battle Sexes game as shown in Table-20.14. Let aand f3
denote the profile of mixed strategies of the Husband and Wife. Which mean
to say that, Husband is playing the F and M with probabilities a and(1 - a)
and the Wife is playing the F and M with probabilitiesf3 and (1 - f3).Let
TCH (., .) and TCw (. , .) be the payoff functions of the Husband and Wife,
respectively'. Thus, TCH(F, F) denotes the payoff of Husband when both the
Husband and Wife are going for Football. Similarly, TCH(M, F) denotes the
payoff of Husband when Husband going for Football and Wife going for
Movie. Similarly, you can interpret TCw(F, M) and TCw(F, F).
The Battle of Sexes game has 4 possible outcomes (F, F), (F, M); (M, F), (M,
M).
We will now calculate the. expected payoffs of Husbands..L, f3), for playing,
the.pure strategies (F) and '(M) under the assumption that the Wife is playing
the Mixed Strategy. Thus, CH(F, f3) = Expected payoff of Husband for
playing the pure strategies (F), under the assumption that the Wife is playing
the Mixed Strategy f3, (1 - f3). Thus,
cH(F,f3) = f3 TCH(F, F) + (1- f3) TCH(F, M)=f3. 3 + (1 - f3). 0 = 3f3 .. (20.9)
CH(M, f3)=Expected payoff of Husband for playing the pure strategies (M),
under the assumption that the Wife is playing the Mixed Strategy f3, (1 - f3).
Thus,
CH(M, f3) = f3 TCH(M, F) + (1 - f3) TCH(M, M)= f3.0 + (1"'" f3).1 = (1- f3)
. (20.10)
Similarly, we will now calculate the expected payoffs of the Wife
cw(F, a), for playing the pure strategies (F) and (M) under the assumption
30 that the Husband is playing the Mixed Strategy with probabilities a, (1 - a).
Thus, Ew(F, a) = Expected payoff of Wife for playing the pure strategies (F), Static Games of
under the assumption that the Husband is playing the Mixed -Strategy Complete
Information
a, (1- a).
That is, Ew(a, F) = a 1fw(F, F) + (1 - a) 1fH(M, F) = a .1 + +(1 - a). 0=
a (20.11)
Ew(M, a) = Expected payoff of Wife for playing the pure strategies (M),
under the assumption that the Husband is playing the Mixed Strategy a, (1-
a). So,
Ew(a, M) = a 1fw(F, M) + (1 - a) 1fH(M, M)= a.O + +(1 - a). 3 =
3(1 - a) (20.12)
Table 20.14: Battle of Sexes under Mixed Strategy
Wife
Football (F) Movie(M)
p (1- P)
Football 1fH(F, F), 1fw(F, F) 1fH(F, M), 1fw(F, M)
Husband (F) , = (3, 1) = (0, 0)
(a)
Movie 1fH(M, F), 1fw(M, F) 1fH(M, M), 1fw(M, M)
(M)' = (0, 0) = (1, 3)
(1~ a)
Under the Mixed Strategy the expected payoffs of Husband from F and M
should be equal (if the expected payoffs are not equal then there can't be
mixed strategy equilibrium).
Therefore equating equation (i) and (ii) for Husband. Thus,
EH(F,f3) = EH(M,f3)
1 3
~ 3f3 = (1- f3) ~ f3 = 4~ (1- f3) = 4:·········· (20.13)
Repeating the same for the Wife, by equating (iii) and (iv) we get:
EW(a, F) = Ew(a,M)
3 1
~ a = 3(1 - a~ ~ a = "4 ~ (1 - a) = "4 (20.14)
Non-Cooperative Plugging the values of a, (1- a), and {3,(1- {3),we can derive the
Game Theory-I numerical values of the Expected payoffs of the Husband and Wife from
equations (20.9), (20.10), (20.11) and (20.12).
From (i) expected payoff of the Husband from the pure strategy F:cH(F, {3) =
3{3= 3). = ~
4 4
From (ii) expected payoff of the Husband from the pure strategy
M:CH(M,{3) = (I - {3) = ~
From (iii) expected payoff of the Wife from the pure strategy F:cw(a, F) =
3
a=-
4
From (iv) expected payoff of the Wife from the pure strategy M:cw(a, M) =
3(1- a) = ~
4
The above values of expected payoffs of the Husband and Wife are derived
for their pure strategies of F and M when the opponent is playing the mixed
strategies.
Now let us consider that both Husband and Wife are playing the mixed
strategies. The expected payoff of the Husband from the mixed strategies is
CH(a,{3) = a.cH(F,{3)+ (1- a)cH(M,{3)
= a{{3 TrH(F, F) + (1 - {3)TrH(F, M)} + (1 - a){ {3TrH(M, F) +
(1 - {3)TrH(M, M)} (20.15)
Thus in .the equilibrium the expected payoffs of both the Husband and Wife
when both of them are playing a mixed strategy are equalized. Moreover this
value is equal to the payoff of the Husband or Wife under pure strategy when
the opponent is playing mixed strategy at the equilibrium (a*,{3*).==>
2)1 f CH(M, {3) > CH(F, {3) ==> Husband's unique best response is to play
pure strategy M.
3) If cH(F,{3) = cH(M,{3) ==> All mixed strategies of the Husband yields
the same expected payoff and hence all are the best responses. Same way
you can do it for the Wife.
32
/
For the ~esp~nse function of the Husband: Static Games of
Complete
1 .
1) For fJ = -4 ==>cH(F,fJ) = cH(M,fJ) ==>Exp. Payoff of Husband from all Information
mixed strategy givesher the same payoff. Thus every mixed strategy is
his best response.
2) If fJ < ~4 ==>cH(F,fJ) < cH(M,fJ) ==>Exp. Payoff of Husband from M
exceeds his expected payoff from all mixed strategy that assign positive
probability to F.==>Unique best response for Husband is M.
1
3) If fJ > - ==>cH(F,fJ) > cH(M,fJ) ==>Expected payoff of Husband from
4
F exceeds his expected payoff from all mixed strategy that assign positive
probability to M. That impliesunique best response for Husband is F.
So, the Husband's best response function is:
1
0, and fJ < 4"
1
a: 05: a 5: l,andfJ = 4"
1
1, fJ > 4"
Similarly, Wife's best response is:
Proposition: Every strategic game in which each player has finitely many
strategies has a mixed strategy Nash equilibrium.
fJ
/1\
(1,1)
1
1
4 ~------------------~-------
a
-
3
(0,0) 1
4
33
Non-Cooperative Recall in the mixed strategy equilibrium we have assumed CH(F, P) =
Game Theory-I
CH(M, p)and cw(a, M) = cw(a, F)in the equilibrium.
Proof: Let us assume statement 1 and 2 hold.
1) The Husband is playing the mixed strategy with probabilities a and (1 -
a)for F and M respectively (0< a<I).
2) But the expected payoff of the husband from F and M (when Wife is
playing the mixed strategy) are not equal i.e., cH(F, P) =1= CH(M, P).
Let us assume cH(F,P) > cH(M,P)
Now from 20.15, we know, the expected payoff of.the Husband when both of
them are playing the mixed strategies is CH(a,p) = a,cH(F,P)+ (1-
a)cH(M,P)
Differentiating the above equation with respect to a (the probability with
which husband plays the F):8e~~,p)=1.cH(F,P)(-1),cH(M,P) =
CH(F, P) - cH(M, P) >0 (because of the assumptionj=e 8e~~,p) >0 ===>
Expected payoff of the Husband from the mixed strategy is increasing in c, As
increases the expected payoff of the Husband from the mixed strategy i.e.,
CH(a,p) increases. So, it is optimal for the husband to put a = 1.
Therefore, the husband can maximize his expected payoff CH(a, P)by playing
F with probability 1 and Movie with probability O.
Hence the contradiction to the statement 1 and in equilibrium CH(F, P) =
CH(M,P)·
Similarly, we can show,
8eH(a,p)
if CH()F,p < CH(M,P ) ===> 8a
1. cH(F, P)( -l),cH(M, P)= cH(F, P) - CH(M, P) < 0 (because of the
assumptionj=e 8eH(a,p) < 0 ===>
8a
Expected payoff of the Husband from the mixed strategy is decreasing in u,
===> As a increases CH(a, P) falls. So it is optimal for the husband to put a = O.
Thereforethe husband can maximize his expected payoff CH(a, P)by playing F
with probability 0 and M with probability 1.
Hence the contradiction. So in equilibrium.sj, (F, P) = CH(M, p)(proved)
Example: The following matrix is a partially specified payoff matrix of the
two player strategic game is given below. Both the players are playing mixed
strategy. Figures in the bracket against each strategy are showing the
probabilities of their mixed strategy. This example is illustrating a method of
checking whether a mixed strategy profile is a mixed strategy equilibrium.
Player 2
(a,p) L(P = 0) C (.!.) R (~)
3 3
3
Player 1 T(a =-) (.,2) (3,3) (1, 1)
4
34
r
Static Games of
So we have to check the indicated pair of mixed strategy {(~, 0, ~), (0, ~,~)} is
4 4 3 3 Complete'
a Mixed Strategy Nash equilibrium. This also suffices by the properties of Information
Mixed Strategy equilibrium. Note that Husband is playing T and B with
positive weights and Wife is playing C and R with positive weights. Consider
the player 1's expected payoff from 3 pure strategies (T, M, B) provided
player 2 plays mixed strategy: '
125
1) c1(T,[J) = O. (.) +-3
333
+ 1.-=-
2) Cl (M, [J) = O. (.) + -31 .0 + -23. .2 = -
4
3
Expected payoff of the Husband when both Husband and Wife are playing
mixed strategy:
Cl (a, [J) = ~C1(T,[J) + O. cl (M, [J) + ~C1(B, [J) = (~. i)+ (0. ~) + ~·i i
=
.........................................................................................
........................................................................................
........................................................................................
Player 2
Tough (t) Concede (c)
Player 1 Tough (t) (1, -1) (10,0)
Concede (c) (0, 10) (5,5)
35
Non-Cooperative
Consider the Mixed Strategy in which Player 1 plays t with probability+,
Game Theory-I . 11
while Player 2 plays t for sure. Then what is the player 1's expected
payoff?
Player 2
Player 1 Left (L) Middle (M) Right (R)
Up (U) (1,0) (1,2) (0, 1)
Down (D) (0,3) (0, 1) (2,0)
Player 2
L R
Player! T (2, 1) (0,2)
B (1,2) (3,0)
Non-Cooperative Core: Set of feasible allocations that cannot be improved upon by a subset (a
Game Theory-I coalition) of the economy's consumers
Strategic Form: A game in strategic form, also called normal form, is a
compact representation of a game in which players simultaneously choose
their strategies. The resulting payoffs are presented in a table with a cell for
each strategy combination.
Strategy: In a game in strategic form, a strategy is one of the given possible
actions of a player. In an extensive game, a strategy is a complete pian of
choices, one for each decision point of the player. .
Zero-Sum Game: A game is said to be zero-sum if for any outcome, the sum
of the payoffs to all players is zero. In a two-player zero-sum game, one
player's gain is the other player's loss, so their interests are diametrically
opposed.
8) A specific move or action that a player will follow in every possible Static Games of
I attainable situation in a game. Complete
i Information
I 9) A strategy is weakly dominant if, regardless of what any other players
I
I
do, the strategy earns a player a payoff at least as high as any other
I strategy, and, the strategy earns a strictly higher payoff for some profile
I
I of other players' strategies.
I' 10)
.. Player 2
L R
Player! L -1,1 1, -1
R 1, -1 -1,1
11)
I
I •
Player 2
H T
Player! H 1, 1 1, 1
T 1, 1 1, 1
12)
Player 2
H. T
Player! H 1, -1 -1,1
T -1,1 1, -1
13)
Player 2
H T
Player! H 1, 1 1, 1
T 1, 1 1, 1
3) Player 1 has two strategies and player 2 has three. S'l= lup, down} and
S2={left, middle, right}. For player 1, neither up nor down is strictly
dominated. Up is better than down if 2 plays left (since 1>0), but down is
better than up if 2 plays right (since 2>0). For player 2, however, right is
strictly dominated by middle (since 2>1 and 1>0), so player 2 being
rational will not play right.
Thus if player 1 knows that player 2 is rational then player 1 can eliminate
right from player 2's strategy space. So, if player 1 knows that player 2 is
rational then player 1 can play the game as if it was the game depicted below.
39
Non-Cooperative
Player 2
Game Theory-I
left middle
Player 1 up 1,0 1,2
down 0,3 0,1
In the figure above, down is strictly dominated by up for player 1 and so if
player 1 is rational (and player 1 knows that player 2 is rational, so that the
second game applies) then player 1 will not play down. Consequently, if
player 2 knows that player 1 is rational and player 2 knows that player 1
knows that player 2 is rational (so that player 2 knows that the second game
applies) then player 2 can eliminate down from player l's strategy space,
leaving the game looking like below.
Player 2
left middle
player 1 Up 1,0 1,2
And now left is strictly dominated by middle for player 2, leaving (up,
middle) as the outcome of the game. This is process is called the iterated
elimination of strictly dominated strategies.
3) Hint: When player 2 plays t for sure, player l' s expected payoff: ~(-1)
. 11
9 9
+ (1--).0
11 = --
11 When player 2 plays c for sure, then player l's
9 9 100
expected payoff: 11(10)+ (1-11).5= 11
20.13 EXERCISES
1) An employee (player 1) who works for a boss (player 2) can either work
(Q) or shirkrf'), while his boss can either monitor the employee (M) or
ignore him (I). Like most employee-boss relationships, if the employee is
working then the boss prefers not to monitor, but if the boss is not
monitoring then the employee prefers to shirk. The game is represented in
the following matrix:
40
Static Games of
Player 2 Complete
Information
M I
Player! W 1, 1 1,2
S 0,2 2,1
_ q), or q > ~, and V2(P, M) > V2(P,I) if and only if p + 2(1 - p) > 2p + (1 -
2
1
p), or p <-. It follows that for player 1,
2
p= if q<-1
.
° 1
2
1
pE[O,I] if q =-
2
1
p= 1 if q>-
2
q= 1 ·f p<-1
1
2
qE[O,l] if p =~
2
if P>21
q=
° 1
Answer: (b)
Best response corresponds to Nash equilibrium (p, q) = G, D
2) Use the rninirnax method to get the mixed strategy of the game.
Solve the following game:
B
A Q ~J
Solution
It is a 2 x 2 matrix and no saddle point exists. So the solution is
!J !
2
y
x 4,1
2,1
2,1
5,0
0,2
0,0
1,
0,3
°
Z 0,0 1, ° 4,1 0,0
a) Compute the set of mixed strategies.
b) Compute the set of all Nash equilibria.
Ans. (a)
w is dominated by a mixed strategy putting 2/3 on x and 1/3 on z.
a is dominated by mixed strategy putting 3/5 on c and 2/5 on d.
b is dominated by mixed strategy putting 3/5 on c and 2/5 on d.
y is dominated by a mixed strategy putting Y2 on x and Y2 on z
d is dominated by c.
x is dominated by z.
The set of rationalizable strategies is {z} x {c}
b) The only Nash equilbrium is (z,c) because there is only 1 mixed
strategy for each player
4) Let PI and P2 be the prices of a homogeneous product produced by firms
1 and 2 respectively. Each firm chooses their price Pi E [0, 00) . There is
no restriction on price and the firms choose their price simultaneously. Di
(Pi, pj) is the demand function of the ithfirm,
1
- F(p.) ... ···if P,
2 I I
= P, J
° ······ifpj> Pj
Both the firms have a constant marginal cost, c. The payoff function of the ith
firm is given by 7lj (Pi' Pj) = (Pi - c) D, (Pi' Pj) •
i) Case I Pi* > P j * > c ~ In this case the entire market is served by the Jh
firm. If the ith firm reduces its price slightly such that its price is below
the Jh firm but above the marginal cost, then it will get the entire market.
Therefore, Pi* is not the best response of the ith firm and hence such a
price configuration cannot be the N ash equilibrium.
42
r
•
. ii) Case 11 P, * > Pi* > c -+ Here the t firm is not behaving optimally (the Static Games of
Complete
argument is the same as case I). Information
iii) Case IIIPj* = Pi* > c -+ Given the equilibrium price of one firm the
other is not behaving optimally. Reducing price by ; (where ; is
negligibly small positive quantity) one firm can increase its profit given
the other firm is charging its equilibrium price.
iv) Case IV p/ > Pi* = C -+ Firm i is not doing its best. It can increase its
profit by charging a price, which is greater than c but slightly, lower than
the pricepj * .
5) Solve the following normal form game using the method of iterated
elimination of strictly dominated strategies.
Player 2
Left Middle Right
Player 1 Up (1,0) (1,2) (0, 1)
Down (0,3) (0, 1) (2,0)
Player 2
Left Middle
Player 1 Up (1,0) (1,2)
Down (0,3) (0, 1)
Notice that in the reduced table for player 1, the strategy "Down" is
strictly dominated by the strategy "Up". Therefore, we can remove it
from the table. The further reduced table is as the following:
Player 2
Left Middle
Player 1
Up (1,0) (1,2)
Now, there is a single strategy left for player! and for player 2 the
strategy "Left" is strictly dominated by "Middle". Therefore, the
equilibrium is (Up, Middle)
6) Can we solve the following normal form game using the mini-max
method? If possible then solve it.
43
Non-Cooperative
Game Theory-I
Player 2
Left Middle Right
Up 5, 1 3,3 1,5
Player 1 Straight 6,0 2,4 1, 5
Down 1,5 0,6 0,6
Player 2
L Ml M2 R
Player 1
U
M
1,
2,4
° 4,2
2,0
2,4
2,2
3,1
2,1
D 4,2 1,4 2,0 3,1
Player 2
Left (L) Centre (C) Right (R)
Top (T) (0,4) (4,0) (5,3)
Player I, Middle (M) (4,0) (0,4) (5,3)
Bottom (B) (3,5) (3,5) (6,6)
44
UNIT 21 STATIC GAMES WITH
COMPLETE INFORMATION:
APPLICATIONS
Structure
21.0 Objectives
21.1 Introduction
21.2 Game Theoretic Applications in Common Property Resources
21.2.1 Prisoners' Dilemma Resource Conservation Game
21.2.2 Chicken Game
21.2.2.1 Problem of the CPR Provision - Two Person Chicken Game
21.2.2.2· Chicken Game: Heterogeneous Group and Collective Action
21.2.3 Coordination Game in Fishing
21.0 OBJECTIVES
After going through this unit; you will be able to:
• understand the nature of strategic interactions among the agents in using
Common Property Resources;
• appreciate application of three different categories of normal form games .
viz., Prisoners' Dilemma Game, Chicken Game and Coordination Game
(already introduced in the Unit 20) in common property resource (CPR)
Management;
• derive best response functions for computation of Nash equilibrium in the
case when payoff functions are continuous; and
• learn derivation of N ash equilibrium in the case of Coumot Duopoly
(quantity competition), Bertrand Duopoly (Price competition), War of
Attrition Game and Hotelling's Linear City Model.
45
Non-Cooperative
Game Theory-I 21.1 INTRODUCTION
Game theory is a multi-person decision theory that analyses situations in with
payoffs to agents. It involves the analysis of conflict, cooperation, and (tacit)
communication and extends applications into several fields, such as
economics, politics, law, biology, and computer science. Quite a number of
examples under discrete payoff structure and continuous payoff structure are
presented here to analyse Nash equilibrium in pure strategy. These
applications help learn theoretical insights, to understand how to formalize
game theoretic problem from an informal description of multi-person decision
making.
46
Table 21.1: Prisoners' Dilemma Game Static Games with
Complete Information:
Player 2 Application
But beyond a certain level grazing, the pasture can no more be sustainable. If
the villagers desire to maintain its sustainability, the number of cattle grazing
there must be restraint so that the extraction of the resources does not exceed
the regeneration capacity of the pasture. When such a participation occurs in
the conservation of the commons, it results in payoff (s, s)given in Table
21.1.
However, if any of the participants them defects from participation (given the
other player participates) then defector's payoff increases and the one who is
participating gets nothing. That is,(p, d) = (v, r) or,(d, p) = (r, v). So
there is always an incentive among the villagers to free-ride or overgraze
(defect).
Thus, mutual defection or free riding (d, d) becomes the dominant strategy
equilibrium. The strategy, defection, for each of the villager dominates the
strategy participation, irrespective of the other players' choice of action (or
participation is dominated by the strategy defection for each). This sort of
outcome is described by Mancur Olson (1965) as - individual rationality
produces collective irrationality.
Here the cooperation is not sustainable as each of the villagers faces a credible
threat that given he participates, his rivals may defect which will push him to
zero payoff. Such a mutual threat pushes them to play safe by choosing
defect (d, d).The outcome is unique pure strategy Nash equilibrium (d, d),
with payoff =(t, t). This strategy (d, d) is the mutually best response strategy
or Nash equilibrium although it is not social / Pareto optimal and both the
villagers over exploit the pasture.
Basically, lack of institutional control may lead the individuals to over exploit
the natural resource base. So long as there are multiple users competing for
the resource, and pay-off structure follows inequality, r > s > t > v,
extraction by one creates negative externalities on others causing Hardin's
tragedy of the commons.
Non-Cooperative 21.2.2.1 Problem of the CPR Provision - Two Person Chicken Game
Game Theory-!
Consider a situation when everybody agrees something has to be done, but the
problem is who will actually do it?
There may be many situations where collective problem is solved through
unilateral actions. The situation arises, when collective goods can be obtained
at sufficiently low costs in relation to its benefits. This may create an incentive
to the bigger stake holders to undertake the responsibility of procuring it.
Given the total gains associated with the production of collective goods or
avoidance of collective bad is pretty large in relation to individual's costs
share in avoiding it, unilateral contribution may be the Nash equilibrium.
Consider two small scale marine fishermen A and B, who stands threatened
by the invasion of a fleet of foreign trawlers in their traditional fishing zone.
Their income is presently 10 units. But to maintain the level of present catches
they need to have inshore water legally protected against the encroachments
of foreign boats. They can secure the legal protection from the government by
making a lobbying effort (which is a public good), which costs altogether, say
Rs.4.
Table 21.2: Fishermen Game
Fisherman 2
Cooperation (C) Defection (D)
Cooperation (C) (x, x ) = (8, 8) (a, b) = (6, 10)
Fisherman 1 (NE)
'Defection (D) (b, a) = (10, 6) (y, y) = (2, 2)
(NE)
The lobbying is so rewarding that it pays a single agent to bear the entire
costs. The agents without lobbying effort (D, D) will face a disastrous
situation in their payoff (y = 2, Y = 2). So there are two-pure strategy Nash
. equilibrium (C,D)or (D, C). Politically active agent (or sucker) will
unilaterally bear the entire costs of procuring the collective goods.
In this kind of game the players have no dominant strategy (unlike the PD
game). As we have seen in PD game, each player prefers others to rake the
responsibility of procuring all goods, while he refrains from moving. However
contrary to the PD games, in this kind of game, each player is willing to take
full responsibility for that action if the other refrains to do anything. The
consequence of nobody doing the action is so disastrous that either of the
players would undertake it if the other refuses to do it.
Other Applications of Chicken Game are:
• Maintenance and surveillance of irrigation system
• Maintenance and surveillance of pastures
• Maintenance and surveillance of grazing fields
• Maintenance and surveillance of hunting grounds
48
r
In some fishing communmes the problem of conflict of appropriation of Static Games with
public goods is solved by reorganizing the claim of the first entrants into the Complete Information:
Application
fishing grounds. There may be some conflicting situations when ill case of
catching of lobsters or bottom dwelling species, the number of fishermen
exceeds the number of locations, where enough lobsters can be caught to
justify the effort involved.
Poor Herder
Cooperation (C) Defection (D)
Poor Cooperation (C) (x, x) = (15, 3) (13,5) (NE)
Herder
Defection (D) (17,0) (NE) (2,0)
If we assume that the rich is more powerful in terms of economic and political
power, it is more likely that class has abetter ability to take prior moves than
the poorer or weaker party. If this is true, then the appropriate tool for analysis
of decision problem involving rich and poor is not the normal form or
strategic of game but the extensive or sequential form of chicken game (which
will be introduced in the Unit 22). The solution concept in that case is
subgame perfect Nash equilibrium.
3) Why the players choose strategies that are not Pareto optimal in common
property of grazing field.
......................................................................................
50
5) Why Chicken game is called anti-coordination game? Static Games with
Complete Information:
Application
on'
_1
oaj
= (c - aj + aa - aj = 0 (21.2)
51
a2rr. a2rr·
Non-Cooperative
Game Theory-I
The second order condition (SaC): aa.;
!
= -2 ai < 0 andaa.;
1
= -2 aj <0
a rr· = 1 > O. So the effort levels a.anda, are strategic complements.
2
andaa.a~.
! 1
Solving the two best response functions (equation 21.1 and 21.2) we get:
c+a'
b.l ( a.J ) = a.l = --land
2
The solution or the Nash equilibrium of this example has obtained the best
response functions such that, at = aj = c .
a~
l
1
-c
2
1
-c a~
2 J
With entry and exit' of the firms are restricted, let us consider a duopoly
market structure with two symmetric firms, summarized by their demand
function and identical cost function.
Each firm's strategy space can be defmed as: Si = [0, oc), output is
continuously divisible and negative output is not feasible.
Thus the payoff function of the 2 firms: 1f1(qV q2), 1f2(qV q2)
There is interdependence in each firm's Payoff functions. Payoff functions are
simply the profit functions and are realized by a combination of strategies
chosen by competing firms. If qi increases then 1fi(') also increases till a
certain point. However, if q, increases, then the profit of the rival firm 1fl.)
falls. So if both firms increase their output, aggregate profit may as well fall.
In order to derive the Nash equilibrium output (q;, q;) we have to derive the
best response functions (reaction function) of both the firms,
Consider the profit function ofthe firm 1: 1f1(qV q2) = p(Q). q1 - c1q1
53
Non-Cooperative Thus, 1fl(qVqZ) = ql(a - bq, - bqz) - clql'
Game Theory-I
Maximizing the profit function of the firm 1 with respect to.qv we get the
Rz (ql) is the reaction function of the firm 2 or, the best response function of
the firm 2. RZ(ql)is a linear function, decreasing in (ql), with slope aq2 = ~
aql
C q2 q l
- ~ 0 and the intercept = (a- 2) > O. (Given that aa = -.!. or, aa = - 2).
Z Zb ql Z q2
Thus, both the reaction functions are downward sloping straight lines. If both
are plotted on the (qz, ql) plane, RZ(ql) is steeper than Rl (qz) ~
a 1 IR1(qZ) = -.!.and aq1 IRz(ql) = -2. If the rival's output increases, it is
q
a q2 Z a q2
optimal for the incumbent firm to reduce the output. The intuition is quite
simple, i.e., if the ith firm increjses its output (for a given level of output of
the rival), price will fall. In order to maintain the high price the rival firm will
have to drop its output. Taking the second order cross partial derivative of the
two firms' profit function we get a2;1(~Vq2) -i < O. Thus, qlandqz are
q2 ql .
strategic substitutes of each other. As the firm 2 expands its outputq-,
marginal profit ~rrl~:~q2) of the firm 1 falls. Same is true for firm 2. In such.
case the reaction functions are always negatively sloped. However, if the
second order cross partial derivative is positive, then ql andqz are strategic
complements of each other. In that case reaction functions are positively
sloped.
Given the reaction functions of firms 1 and 2, we have a unique intersection
which defines the Cournot Nash equilibrium. Solving the reaction functions
(equation 21.6 and 21.7) corresponds to the intersection of the functions.
Thus,
a-ZCl + C2
*
ql = 3b ' qz
*
= a +Cl3b- C2 .
.•.......... (21.8)
d-q
b
............ (21.11a)
I will increase and that of the firm to will fall (you may check olri~qi,qi) <0
Cl
011'* ( * *)
and 2 %q2 > 0). See Figure 21.3.
OCl
55
r
Non-Cooperative
Game Theory-I
q2M ql I =0 -
- (a-c) -
Zb - qM, ( 2 1.1 3)
[Note: You may check the equilibrium monopoly output under the linear
demand function p(q) = a - bq, and constant marginal cost C(q) = cq. The
profit function ofthe monopolist is rrM = pq - cq = (a - bq)q - cq] .
M
FOC of the profit maximization of the monopolist: d1l = (a - 2bq - c) =
dq
o~ qM = (a-c)
2b
is the equilibrium monopoly output. The monopoly profit
(a-c)2
W ill b e=--.
4b
Suppose the two symmetric duopoly firms agree to collude together into a
single entity to take output decision jointly and thereby share the monopoly
profit equally. Then the colluded firms behave just like a monopolist. If the
marginal costs are constant, any division of monopoly output qM (any
division of profit) between the 2-frrms will give industry profit equal to the
monopoly profit. All possible division of monopoly output between the two
duopolist is shown by the line segment qtt q~. Any equal division of
monopoly output and profit corresponds to point M. (See Fig 21.3).
We have seen under symmetric condition,Coumot equilibrium output (from
equation 21.12)isqf* = qi* = qC = (a-c).
3b
Each firm would have a profit:
rr*c - rr*c - rr*c _ (a-c)2
1 - 2 - - 9b
56
Hence Static Games with
M _ (a....•
c) > c _ (a-c) Complete Information:
q - 2b q - 3b ==> Application
Monopoly output is greater than Cournot output at the firm level.
h . qid = R ~(qf)
were 2 ,1,. J. = 1, 2 . P'oint M (qM
2' .qM) .
2 ISnot th e equui
ilibnum
. . IS
as It .
not profit maximizing for any of the firm, This is because it is not in the
reaction function (or, the best response function) of any of the firm. The
collusive agreement is not sustainable as it is not a Nash equilibrium. See
Figure 21.4.
ql
2b
qM
2
c
qM q2
2 2b b
q~
Fig. 21.4: Cournot vs. Collusion
57
r
Non-Cooperative Duopolists will be better off if they collude. But given the other firm
Game Theory-I producing the collusive output, each has an incentive to deviate. The firm
which respects the agreement will be worse off (compared to its Coumot
outcome) while the firm which ditches the partner firm by deviating will be
better off. So there is a credible threat for cooperation for each of the firm
because the other may ditch. As a result both the firms are worse off when
each responds to its own private incentives. Hence a Coumotduopolist faces a
situation of Prisoner's Dilemma.
Solving equation (a) and (b), Coumot Nash equilibrium output of each firm is
qf* = q~* = qC = 40.
Therefore, the Coumot industry level output: QC* = qf* + q~* = 40 + 40 =
80
Equilibrium Coumot Price=p(Qc*) = 130 - QC* = 130 - 80 = 50
Coumot firm level profit of the furml =
1lf*(qf*, qi*) = p(Qc*)qf* - cqf*= 40(130 -80-10) = 40.40 = 1600.
Similarly, equilibrium profit ofthe firm 2=1l~*(qf*, qi*) = 40.40 = 1600
We get the industry level profit as 1600 + 1600 = 3200.
Now consider the two firms collude together by jointly producing the
monopoly level of output and share the profit equally.
Monopoly profit (or, collusive profit) of the industry=r'" = (130 - qM)qM -
10qM
Monopoly output (or, collusive output) of the -industry can be obtained by
maximizing rrMw.r.t. qM.TheFOC of the profIt maximization of the
drrM
monopolist=- = (130 - 2qM - 10) = 0 ~ 2qM = 120 ~ qM = 60.
dQ
58
M M Static Games with
So the collusive output of the each fIrm(L ,L) = (30,30). Complete Information:
2 2
Application
Equilibrium price: pM = 130 - qM = 130 - 60 = 70.
Equilibrium profit under collusion-er" = (130 - qM)qM - 10qM =
M M
60(70 - 10) = 60.60 = 3600. Given the firms share the output (L, L), the
2 2
M M
individual firm share of the profit: (~,~) = (1800,1800). Thus the firms
2 2
. nM nM
are better off by colluding because they get higher profit share: (-,-) =
2 2
(1800,1800) > (nf*, ni*) = (1600,1600). That means, each firm has an
incentive to collude.
However, each firm has an incentive to. deviate given the partner firm is
M
committing to collusion by producing L.
2
I• Thus, the output of the firm 1 under defection (when firm 2 sticks to collusion
. qM qM 1
59
r
NOD-Cooperative
Game Theory-I
Rationing Rule: Di(Pl ,P2 ) =
2) If Pi = candp, > c i. e., Pi < Pj ,then firm i is charging lower price than
firm j. The consumers will be purchasing from firm i only. So the entire
market will be captured by firm i and firm j will be out competed
i. e. , qj = O.Then firm j can slash the price till Pj = c in order to capture
the market share.
3) If Pi = Pj' then the market will be equally shared by the two firms.
4) If Pi > c, Pj > c, and Pi > Pj thenthe entire market will be captured by
firm j and firm i's market share will bezero. But the firm j is earning
positive profit since Pj > c. In that case firm i can reduce the price below
Pj in order to capture the entire market. This price undercutting will
continue till Pi = Pj = c.
Profit or payoff function of [rrm i under the rationing rule is
(a - Pi)(Pi - c), andif Pi < Pj,Pi ~ a
a-pi .
Tri(Pl, P2 ) = (-2-)(Pi - C), andif Pi = Pj,Pi ~ a
{
. 0 if Pi > Pj
In contrast to the Cournot game (where the profit function is continuous with
respect to the strategic variable i.e., quantity), there is a discontinuity feature
in the payoff functions under Bertrand price competition. The profit of a firm
can sharply drops to zero for charging a very little higher price. A slight
undercutting would enable it to grab 100% market share.
pM,
Now we have to derive the best response function for the firm i and j, i.e., we
have to derive those strategies which give the firm i (and firm j) maximum
payoffs.
1) If Pj < c, ~ Firm i's profit is negative if Pi :::;Pj and zero if Pi > Pj'
So any price Pi > Pj is the best response to Pj· ~ Bi(pj) = {Pi: Pi >
Pj}
Thus in Figure 21.5 (a), the dark segment of the firm i's profit function
shows the relevant profit, whenp, < c. When Pi = Pj < c, both the
form will earn negative profit, as the market' will be equally shared
between the two frrms. So ith firm profit is a discrete point shown in the
• . I
negative aXIS.
2) If Pj = C, then analysis is similar to the previous case except that
Pi ~ Pt yields Tri = 0 and hence B, (pj) = {Pi: Pi ~ Pj}
3) If < Pj :::;Pm,then Firm i can set price Pi :::;P! and earn positive profit.
Tri(Pl pj)is increasing in Pi till Pi :::;Pj and Tri(Pi Pj) = 0 at Pi > Pj·
I I
61
Non-Cooperative See Figure 21.5(b), the dark segment of the firm i's profit function shows
Game Theory-I that for a given Pi : if the firm i chargesp, < c, it will capture the entire
market but earn negative profit. For any Pi : c < Pi < Pj , firm i will earn
positive profit and enjoy entire market. When Pi = Pj ,the market will
be shared between the two firm, so the firm i's profIt turned out to be a
discrete point, For any price Pi > Pj , firm i will loose the entire market
share and earn zero profit,
4) If Pj > Pm, then charging the monopoly price (Pm) is the unique best
response of firm ii. e., Bi(pj ) = Pm'
See Figure 21.5( c), the dark segment of the firm i 's profit function shows that
for a given Pj > Pm, for any price less than Pj .firm i will enjoy the entire
market share and is increasing in Pi' For Pi = Pj' firm i's profIt drop to a
discrete point and becomes zero for any price Pi > Pj'
Thus firm i's best response function is
{Pi: Pi > Pj}, andif Pj < C
pM
c ----- c
" "II " "II
""
" 45° I
I
""
" 45°
I
I
0
" C
" I
pM Pj
0 C pM
Pj
62
r
Static Games with
Check Your Progress 3
Complete Information:
Application
1) What do you mean by tacit collusion?
.....................................................................................
......................................................................................
......................................................................................
.
.....................................................................................
.
......................................................................................
2) What does Bertrand paradox refer to?
.....................................................................................
......................................................................................
.....................................................................................
.....................................................................................
.....................................................................................
3) In a Coumot duopoly when does a Nash equilibrium occur?
.....................................................................................
.....................................................................................
., .
.....................................................................................
.....................................................................................
4) Consider the case of Coumot Duopoly (n = 2) when the market demand
is p = 10 - Q and the firms are symmetric in costs with C(qa =
1. qi, where i = 1,2.' Show that the Cournot equilibrium quantity
(qi' qa = (3,3), equilibrium price p = 4 and profit at the firm
level = 9.
.....................................................................................
......................................................................................
.....................................................................................
.....................................................................................
.....................................................................................
.....................................................................................
5) Consider the case of Coumot Duopoly (n = 2) when the market demand
is p = 10 - Q and the firms are symmetric in costs with C(qi) =
1. qi' where i = 1,2. Show that if both firms colludes and shares the
profit equally, then the equilibrium output of each firm (under
collusion)= (2.25, 2.25), Price = 5.5, Each firm's share of profit = 10.125.
.....................................................................................
.....................................................................................
. ..~ .
.....................................................................................
.....................................................................................
....................................................................................... 63
r
Non-Cooperative 6) Show that by cheating on the Cartel (Collusion) a firm increases its own
Game Theory-I profits but necessarily lowers its riva1' s profit by an even larger amount.
..
..................................................................................... .
A Simple Example
For our understanding the War of Attrition game, let us consider the following
simple example:
Suppose two companies, A and B, must decide each month whether to spend
Rs.1000 lakh. If in the first month one company spends the Rs.1000 lakh and
the other does not, the game is over - the first company becomes a monopolist
worth Rs.10000 lakh, and the second company must exit from the race. If
neither invests Rs.lOOO lakh in the first month, the game is also over, with
neither company losing nor making money. However, if both companies
spend Rs.l 000 lakh in the first month, neither one wins anything.
Now move to the second month, where again each company has to decide
whether to spend Rs.1000 lakh. If both companies again spend Rs.1000 lakh,
the game moves to the third month, and so on. If, at the start of some month,
one of the companies spends Rs.lOOO lakh and the other does not, the first
company wins the Rs.l0000 lakh prize. But of course many months (and
much money) could go by before this happens.
Again, think carefully about these questions before proceeding further. Static Games with
Complete Information:
Initial conceptualization of the gamehas come from strategies related to the Application
conflict between two animals fighting over a prey. Each one chooses the time
at which it intends to give up. Fighting involves efforts. So, it is costly.
Naturally, each animal prefers the fight as short as possible. If both animals
intend to give up at the same time, then each has an equal chance of obtaining
the prey. Let the fighting (or waiting time) is continuous which starts at 0 and
continues infinitely. Also, let there be two players iandj. Each player's set of
action is the set of possible concession times. Let Vi be the ith individual's
value ofthe object of dispute (Vi> 0).
1) If the player i concedes first at the timer., his payoff is (- td, as he
spent ti units of time and does not obtain the object.
2) On the other hand, if the other player concedes first at the timer., player i
gets the object. So ith player's payoff: (Vi - tj).
3) If both the players concede at the same time ti = tjthenplayer i's payoff
1
= -2 (Vi - td·
•
Ok-------_e------
It. = Vj t,
I J
I
I
I
l 65
Non-Cooperative Best response function of the player i (for Vi > vi»
Game Theory-I
(b)
(a) t·J
tj B/tj)
Bj(tj)
Vj
vj
--------- ~
'" . I
'" " I
'" I
.'
I
M •
I
lIo..
,
Fig. 21.8 (a, b) : Player i andj's Best Response Functions (Bi(tj), B/ti).
Superimposing the player i and j's bestresponse functions, we see there are
two areas of intersection: 1. Vertical axis ~ Vi 2. Horizontal axis at and to the
right of Vj.
2. tj = 0, ti ~ Vj
.. , .
. 66 .......................................................................................
r
I
2) What is the easiest way of finding the Nah equilibrium in the war of Static Games with
attrition game? Complete Information:
Application
Initially the rn'odel was developed as a game in which firms first chose a
location and after a selling price for their products. In order to set their
business in the best location to maximize profits, the firms will have to
evaluate three key variables: competitors' location, customers' distribution
and transportation costs. We can consider static model where firms choose
their location and prices simultaneously.
The basic idea is that two ice cream vendors are on a beach that stretches the
0-1 interval. Customers are uniformly distributed along that interval. The
vendors simultaneously select a position. Customers go to the closest vendor
and split themselves evenly if the vendors choose an identical position. Each
vendor wants to maximize its number of customers.
The Nash equilibrium is for both vendors to select the median location (0.5);
doing this guarantees each vendor half the business, but deviating to any other
point generates strictly less.
Consider the case of electoral competition where there are two political parties
1 and 2, viz., (n = 2) contesting for votes. Suppose there exists a linear city [0,
1], where voters are uniformly distributed over the interval [0,1] with a unit
density i.e., each points represents one voter. It is assumed that the political
parties have no political conviction but wants to maximize their vote share by'
strategically locating themselves across the linear city. On the other hand,
voters have the political conviction and therefore the voters try to locate
themselves nearer to their political ideology. What would be the Nash
equilibrium voting configuration?
67
r
Non-Cooperative
Case 1: X1 < X2 *- -1
Game Theory-I 2
0
• •
Xl X2
•-1
2
1
Fig. 21.9 (a)
Party 1 and 2 are located at a distance x1andx2, (X1 < X2 *- ;-) distance
from the origin (0 at the extreme left). All the voters to the left of X1 will will
all vote to Party 1. However the voters to the right of Xv (X1 < X2 ) will be
splitted between Party 1 and 2. Thus, the vote share of Party 1: X1 +
~ (X2 -
2
X1) = ~(
2
X1 + x2). In order to understand whether this
configuration is a Nash equilibrium, we have to justify whether it is a
mutually best response strategy.
Suppose the Party 1 deviates from X1 distance to the right of Xv to say, X1"
where X1' = X1 + E, where E > O. .
Now the vote share of the Party 1 is: X1'~ (X2 - X1 ')=~(X1' + X2)=~( X1 +
. 2 2 2
E+X2) = ~( X1 + X2) + ~ E. SO the Party 1 gains by relocating itself to the
2 2
right. So the Party 1 has an incentive to deviate. Hence this is not the Nash
equilibrium strategy.
Case 2: X1 = X2 *- ~2
o • 1
2
• 1
Figure 21.9 (b)
In this configuration both the parties are located together but not in the centre.
Suppose P is the point where both are located. Each party will have equal vote
share, i.e.,;- x1 +;- (1 - X1) = ;-.
Now if any party gains from deviation we can claim that this is not a Nash
equilibrium. Suppose the Party 1 deviates from X1 distance to the right of
Xv to say X1', where X1' = X1 + E, where E > O. So the Party l' s vote
share =(1-
.
X1') + ~(
2
X1' - X1)'
Let the initial distance of both the parties was less than the centre of the linear
city i. e., (1 - X1') > ~~
2 22.
X1' - X1) >~. So, if any of the
(1 - X1') + ~(
parties deviates (to the right in this example) there are gains from deviation.
Such a configuration is also not the N ash equilibrium.
1
Case 3: X1 = X2 = -
2
•
o 1
Fig. 21.9 (c)
68
Suppose the Party 1 deviates to its left at Xl', where Xl' = Xl - C. Static Games with
Complete Information: .
SO the Party 1's vote share after deviation: Xl' + ~(Xl - Xl) [( Xl - Application
1 1 1 1
c) + z( Xl - Xl + c)]=[(Xl - c)+zc] = (Xl - zC) < z" See that the
deviation by either party is not gainful. Hence Xl = X2 = .!. is the Nash
2
equilibrium. So both the parties would prefer to locate themselves at the
centre in order to maximize their vote share.
Check Your Progress 5
1) Suppose consumers are located uniformly along the line segment from 0
to 1. If firm A and B are located somewhere within the line segment,
discuss the Nash equilibrium you will get.
3) In linear city model of Hotelling two fIrms share - transport costs with
their customers.
70
Static Games with .
21.10 CHECK YOUR PROGRESS Complete Information:
ANSWERSIHINTS Application
21.11 EXERCISES
1) Write a short note on the CPR management dilemma.
2) Derive the Nash equilibrium under the Cournot competition in the case of
3 symmetric firms with constant marginal costs (c). The market demand
function is p(Q) = a - bQ = a - b(ql + qz + Q3)' Derive the
equilibrium price and profit at the firm level as well as industry level.
3) Consider a Bertrand duopoly market structure. Defme Bertrand Paradox.
4) . Derive the best response function of the War of Attrition game. Show
graphically and explain the Nash equilibrium.
5) Discuss the Nash equilibrium under Hotelling's location game in the
context of electoral competition with two political parties.
6) Consider an example of the Tragedy of Commons. Show how the
phenomena of over fishing can be represented as a Prisoners' Dilemma.
You may set up the game with two players, each of which can undertake
low or high fishing activity.
Solution:
The case of over fishing should be set up in a manner similar to this:
India
Low
High fishing fishing
activity activity
V.S High fishing activity 1,1 3,0
Low fishing activity 0,3 2,2
The sustainable fishing catch is higher when both nations undertake low
fishing activity.
However, there is then an incentive for both to increase fishing. In fact,
high fishing is a
dominant strategy for both players. We therefore, end up with the worst
outcome.
7) There is competition that conveys the following information. Prize worth
Rs.l, cost c to fight per period. Each player chooses probability p .of
exiting per period. If both exit, there is a split of the/ prize. Best response
requires %p = p- c + (1 - p) %p. Value of exit = value of staying in.
72
r
Solve the game to get how long the the war would last, expected profits Static Games with
Complete Information:
for competitors.
Application
Solution:
Get = 1 - "';1 - 2c. Thus, C = (1/2)c. That means for c = 0.1, war will
last five years on average.
Expected profits areY2p. As profits are increasingin c, the effect on
competitor outweighs own effect.
_ 120-qB
qA ( qB) - 2
Since the game is symmetric (firms have identical cost functions), the
reaction function of firm B is:
A ( ) _ 120-qA
qB qA - 2
ii) Coumot equilibrium is identified by the quantities that are mutually best
responses for both firms; so, they are obtained by the solution of the
following two-equation system:
73
r
Non-Cooperative
Game Theory-I
Q* = 80and the equilibrium price is
P(Q*) = 140 - 80 = 60
and firms' profits are:
Suppose that two firms both have average variable cost c = $50. Assuming
that fIrms complete in prices, then:
1) Defme the reaction functions of the firms;
2) Find the Bertrand equilibrium;
3) Would your answer change if there were three firms? Why?
Solution
1) The construction of the reaction function in a competition a la Bertrand
proceeds in the following manner.
Since firms compete in prices, we· need to use the direct demand
function where Q is a function ofP. From the inverse demand:
100- P if P < 100
Q(P) = { .
o otherwise
There are two firms, firm 1 and 2. Consider now the effect of the price
choice of any firm I on its own profits for any given price chosen by
firm j with i, j = 1, 2 and i '" j
74
I-
I a) Pi > P, : Il, =0 Static Games With
Complete Information:
In this case, firm 1 sells no output and therefore gets zero profits. Application
This is where firm i shares the total profits in the industry. The total
profits of the industry are here determined in the following way.
Price minus average cost gives the profit per unit of output (Pi - 50)
and this is multiplied by the total output (lOO-Pj).
c) Pi < Pj : lli = [(Pi - 50) (100 - Pj)]
Here firm i gets the whole profits of the industry.
In order to firm the reaction function of firm I we can distinguish 3
different cases. .
a) Pi (Pj) = 50 if P, ~ 50
In such a case, any price Pi < Pj will make negative profits, so [If m i
will prefer to lose the race rather than beating j on price. More
precisely we can .say that [If m i will never set a price below c = 50.
This is usually portrayed as the strategy that if Pj < 50, Pi = 50.
b) Pi (Pj) = Pj - 1::' if 50 < Pj ~ 75 = pM
Where pM is the monopolistic price. If Pj > 50, there is potential for
'positive 'profIts, as long as P, < 100, at which point quantity
demanded falls to zero, so that profits would be zero too. The first
issue is, should firm i match the price of [If m j, or attempt to
undereut its rival? Intuition suggests that the [If m should undercut
its rival rather than match Pj, This can be shown formally by
comparing the profits for [If m i if it matches Pj, lit, with the
profits, n,", it gets if undercuts Pj by a small amount I:: so that Pi =
Pj - 1::.
This means:
A
As I:: -) 0, that is as the Pi, gets closer and closer to Pij, llt -) 2lli : if
firm i sets its price just below Pij' its profit will be roughly double the
profit from matching Pj. However, the fact that [If m i will do better by
just undercutting firm j than it does by matching firm j does not mean that
just undercutting is always the best strategy. When the [If m has the whole
market, its optimal price is the monopoly price. Therefore, when Pij is
greater than the monopoly price, firm j should set the monopoly price. In
this example, the monopoly price pM = 75. Following this discussion:
c)' Pi (Pj) = 75 ifPj > pM
The reaction function of any firm is then the following:
75
Non-Cooperative 50 if PJ-< 50
Game Theory-I
P (P)J
1
= P,J - E if 50 < P,J-< 75
\ 75 if Pj > 75
2) The Nash equilibrium is where the two reaction functions interest, which
is where PI = P2 = 50. In such a case the best response of the two firms
are mutually consistent. This is actually the only point in which this is so.
(Please check this result by your own Can you find another Nash
equilibrium?). This is the same as the competitive solution. Price is set
equal to average (= marginal) cost and zero profits are made.
76
r
Structure
22.0 Objectives
22.1 Introduction
22.2 Extensive- form Representation of Dynamic Games
22.2.1 Strategies in Extensive-form
22.3 Dynamic Games of Complete and Perfect Information
22.3.1 Backward Induction
22.3.2 Strategies in Dynamic Games with Complete Information
22.3.3 Subgaines
22.3.4 Subgame-Perfect Nash Equilibrium
22.3.5 Backward Induction and Subgame Perfection
22.3.6 Application 1: Stackelberg Competition
22.3.7 Application 2: Sequential Bargaining
22,4 Dynamic Games of Imperfect Information
22.4.1 Imperfect Information and Backward Indu~tion
22.4.2 Subgames with Imperfect Information
22.4.3 Strategies with Imperfect Information
22.4.4 Finding SPNE with Imperfect Information
22.0 OBJECTIVES
After going through this unit, you will be able to:
• understand the idea behind dynamic games;
• present the dynamic games in extensive form keeping in view complete,
perfect and imperfect information;
• use backward induction as a solution concept to arrive at subgame perfect
Nash equilibrium; and
• analyse repeated games.
77
Non-Cooperative
Game Theory-I 22.1 INTRODUCTION
We have seen in Unit 20 the presentation ofa normal form game. In the
discussion there players choose their strategies simultaneously, without
knowing the choices of the other players. So these are static games. In this
unit we take up dynamic games by allowing the players have the choice
moving sequentially. That is, players make moves at different times or in turn.
So, one player chooses her action before the others choose theirs. That means
players who move later in the game have additional information about the
actions of other players.
Such a method depicts the tree form of the game. Thus, game trees are made
up of nodes and branches, which are used to represent the sequence of moves
and the available actions, respectively. To appreciate the idea, consider two
players, Black Smith and White Elena, who are playing a sequential game.
Black moves first and has the option of Up or Down. White then observes his
action. Regardless of what Black chooses, she then has the option of High or
Low. The tree form of the game is follows:
actions. In other words, a strategy specifies what action a player will take at
each decision node. Consider the game given above between players Black
and White.
Player Black has two strategies available - Up and Down. Player White,
however, actually has four strategies available since there are two nodes to
consider - b and c- and two possible actions at each node - High and
Low. Thus, strategies available to player White summarised in Table 22.1.
Table 22.1: Strategies of the Player
79
Non-Cooperative 3) What is meant by strategy in an extensive form representation of a game?
Game Theory-I
........................................ ~ : .
80
Dynamic Games with
Complete
(2,2) Information
(1, 1)
Here, player II moves last. Since she knows the play will end after her move,
she can safely select the action which is best for her. If player I has chosen to
provide high quality service, then the customer prefers to buy, since her
resulting payoff of Z is larger than I when not buying. If the provider has
chosen Low, then the customer prefers not to purchase. The choices available
to player II are indicated by arrows in figure.
When the last move is decided, backward induction proceeds to the player
making the next-to-last moves (and the moves continue). In Figure 22.2,
player I makes the next-to-last move, which in this case is the first move in the
game. Being rational, he anticipates the subsequent choices by the customer.
He therefore realizes that his decision between High and Low is effectively
between the outcomes with payoffs(2,2) or (1,1). Clearly, his choice for High
results in a payoff of 2 for him. In case of a selection of Low, the outcome is
payoff 1. So the unique solution to the game, as determined by backward
induction, is that player I offers high quality service, and player II responds by
buying the service.
The solution through backward induction specifies the way the game will be
played. Starting from the root of the tree, play proceeds along a path to an
outcome. Note that the analysis yields more than the choices along the path.
Because backward induc~on looks at every node in the tree, it specifies for 81
Non-Cooperative every player a complete plan of what to do at every point in the game where
Game Theory-I the player can make a move. Note that a player's plan includes a node even
when that point may never arise in the course of play. Such a plan is called a
strategy of the player. For example, a strategy of player II in Figure 22.2 is
"buy if offered high-quality service, don't buy if offered low quality service."
Such a strategy for player II's is obtained by backward inductiop.
With strategies defined as complete plan of move by the player, one can
obtain the strategic form representation of the extensive games. Recall that the
strategic form game tabulates all strategies of the players. In the game tree,
any strategy combination results into an outcome of the game, which can be
determined by tracing out the path of play arising from the players adopting
the strategy combination. The payoffs to the players are then entered into the
corresponding cell in the strategic form. Figure 22.3 shows the strategic form
for quality choice example. The second column is player II's backward
induction strategy, where "buy if offered high-quality service, don't buy if
offered low-quality service" is abbreviated as H: buy, L: don't.
:I'
A game tree can therefore be analyzed In terms of the strategic form. It is not
difficult to see that backward induction defmes a Nash equilibrium. In Figure
22.3, it is the strategy combination (High; H: buy, L: don't).
22.3.3 Subgames
A subgame is a subset of any game that includes an initial node. To
understand subgames use the extensive form representation. In the following
game tree (Figure 22.4) there are six subgames other than the game itself, two
of them containing two subgames each.
82
Dynamic Games with
Complete
Information
PLAYER 1
5,2
0,0
U 2,2
- -5,2 --
2,5 2,5
PLAYER 1
3,1
Player 2
L R
Player 1 A 1,2 1,2
B 0,0 2,1
1
0,0 2,1
This game has two Nash equilibria: (A,L) and (B,R). If the players have
planned to select right move before the game starts, then the NE profile (A,L)
makes sense. If player 2 is convinced that player 1 will play A, then playing L
is a best response. In fact, she would do everything in her power to convince
player 1 that she will play L, so that he plays A.
While player 2 might "threaten" to play L before the game starts, if the game
proceeds and it is actually her turn to move, then L is not rational. The threat
is not credible, and player 1 should realize this and play B.
Thus, L might be a reasonable plan for player 2 if she thinks player 1 will play
A, but it is not rational to carryout the plan after player 1 plays B. It is not
sequentially rational for player 2 to play L.
0,0 2, 1
Fig.' 22.6: Optimal Choice in Extensive-Form Game
Next, examine the smaller games. In Figure 22.7, player 1 can either choose
A and reach the terminal node with payoffs (1,2), or B and reach the terminal
node with payoffs (2,1).
1
1,2
0,0 2, 1
Fig. 22.7: Extensive-Form Game with Optimal Choice BR
We find that player l's optimal choice is B. So, backward induction leads to
the strategy profile (B,R), with payoffs (2,1).
Note that every [mite game of perfect information can be solved using
backward induction. If each player has a strict preference over his possible
terminal node payoffs (no ties), then backward induction gives a unique
sequentially rational strategy profile. The resulting strategy profile is a Nash
equilibrium.
The following points may be useful in while dealing with the solution concept
subgame perfect equilibrium:
• A strategy profile for an extensive-form game is a subgame perfect Nash
equilibrium (SPNE) if it specifies a Nash equilibrium in each of its
subgames.
85
Non-Cooperative • The idea behind SPNE is that even if a NE strategy profile dictates that
Game Theory-I
certain subgames are not reached, we require that what the players would
. do conditional on reaching those subgames should constitute a NE. The
"off-the-equilibrium-path" behaviour can be important, because it affects
the incentives of players to follow the equilibrium.
• Notice that every SPNE must also be a NE, because the full game is also a
subgame.
• For [mite games of perfect information, any back ward induction solution
is a SPNE and vice-versa. The advantage of SPNE is that it can be applied
to games of imperfect information too.
Let us adapt the Cournot model covered earlier. Marginal production cost is
.,. equal to 100, and market inverse demand is given by p = 1000 - q1 - qz .
(If the expressionis negative, we take the price to be zero.)
Firm 2' s strategy set is more complicated, reflecting the extensive-form nature
. of the game. Firm 2 can let q2 depend on q 1.
Taking the derivative with respect to ql, setting it equal to zero, and solving
for q1yields450 - ql = 0
qi = 450.
1) The Model
Take two players, one and two, playing the game sequentially and
making offers about how to divide a pie of size one. Time runs from
t = 0,1,2, ..... At time 0, playerone can propose a split (xo,l -
. xo)(with Xo E [0, 1]), which player 2 canaccept or reject. If player 2
accepts, the. game ends and the pie is consumed. If player two rejects, the
game continues to time t = 1. She now proposes a split(yv 1 - Y1)'
Such a proposal of player two may be accepted or rejected by player
one. The sequence of the game goes to ad infinitum.
Let us assume that both players want a larger slice, and also that they
both dislike delay. Thus, if agreement to split the pie (x, 1 - x)is reached
at timet, the payoff for player one is 8fx and the payoff for player two
is 8~(1 - x), for some 81,82 E (0,1).
2) A Finite-Horizon Version
We consider a variant of the game where there is some [mite number of
offers Nthat can be made. If we solve for the subgame perfect
equilibrium through backward induction, we need to start from the final
offer.
Suppose player one makes an offer at a given date t. then player two.
acceptance of it depends on her belief about what she will get if she rejects.
This in turn depends on what sort of offer player one will accept in the next
period, and so on .. Even with such a time horizon we will show there is a
unique subgame perfect equilibrium in the sequential bargaining game.
When player one proposes, she suggests a split (x,l - x)withx = (1-
82) / (1 - 8182), Player two accepts any division giving her at least 1 - x.
When player two proposes, she suggests a split (y,l - y)with Y = 81 (1 -
82) / (1 - 8182), Player one accepts any division giving her at least y. Thus,
bargaining ends immediately with a split(x, 1 - x).
Now consider a period when player one offers. Player one has no profitable
deviation. She cannot make an acceptable offer that will get her more thanx.
Because when the offer made by her is rejected she will get y = 81x the next
period, or 8f x in present terms, which is worse than x. Player two also has no
profitable deviation. If she accepts, she gets 1 - x. If she rejects, in the next
period she gets 1 - y, or in present terms 82(1 - x) = 82(1 - 81x). A
similar argument applies to periods when player two offers.
We now show that the equilibrium is unique. To do this, let El, 171denote the
lowest and highest payoffs that player one could conceivably get in any
subgame perfect equilibrium starting at a date where he gets to make an offer.
To begin, consider a date where player two makes an offer. Player one will
certainly accept any offer greater than 81171 and reject any offer lessthan 81171'
Thus, starting from a period in which she offers, player two cansecure at least
1 - 81171 by proposing a split (81 Vv 1 - 81171)' On the otherhand, she can
secure at most 1 - 81v1.
Now, consider a period when player one makes an offer. To get player two to
accept, he must offer her at least 82 (1 - 81Vl)tO get agreement.
Thus:
El ~ 1 - 82(1 - 81El)'
.............~ ' .
2) What is backward induction?
....................................................... ' .
3) Define a subgame.
89
Non-Cooperative 5) Explain the relationship between Backward Induction and Subgame
Game Theory-I Perfection .
........................................................................................
........................................................................................
..........................................................................................
........................................................................................
........................................................................................
6) What are the three important components a price leader in the
Stackelberg model is assumed to know?
........................................................................................
........................................................................................
........................................................................................
........................................................................................
........................................................................................
........................................................................................
........................................................ ' .
........................................................................................
........................................................................................
.........................................................................................
9) Use backward induction to determine the subgame perfect equilibrium of
the following game:
A,B
0,2
2, 1
A,B
1,0
........................................................................................
........................................................................................
........................................................................................
90
Dynamic Games with
10) Use backward induction to determine the subgame perfect equilibrium of
Complete
the following game: Information
A,B,C
1, 1, 1
A,B,C
2,3,2
0,0,2
1,2,4
0,2,0
........................................................................................
........................................................................................
. .
........................................................................................
........................................................................................
........... , .
11) Use backward induction to determine the subgame perfect equilibrium of
the following game:
A,B
4,5
A,B
1,0
2,2
........................................................................................
........................................................................................
........................................................................................
........................................................................................
........................................................................................
These games can be represented in extensive (game tree) form by using non-
singleton information sets to indicate what each player knows (and does not
know) when she has to the move. Since each player does not know all earlier
moves, we must partition the set of player's decision nodes into subsets, called
information set. We connect the nodes are connected with a dashed line if the
moving player does not know from which node of the game tree he is making
the move (see Figure 22.8).Sometimes instead of dotted line we can use oval
format to represent information set.
(1,-1) A wins
(-1,1) a wins
(-1,1) a wins
B Bs co
92
Dynamic Games with
22.4.1 Imperfect Information and Backward Induction Complete
In games of imperfect information backward induction procedure has to be Information
considered when
• a player does not know which node he is moving from he cannot single
out the best action;
• preceding player into the game tree cannot anticipate what she will do;
• common knowledge is not able to give the usual information;
• a dominated strategy not available.
A special case: prisoner dilemma
3,3
Bs
coo e •I
<,;
A
.... ~:~ : •
BI~
2,2
In tree above given asa special case of prisoner dilemma, even if B does not
know if she is in as
or in BI she can observe that, whatever her starting node,
"defect" gives her more utility then "cooperate":
• 4 instead of 3 if A has done "cooperate";
• 2 instead of I if A has chosen "defect" This is because "cooperate" is a
dominated strategy.
93
Non-Cooperative
3,3
Game Theory-I
I
A •
I .1,4
• 4,1
l~
B~
In the figure player A has one information set that is a singleton. Player B has
one information set with two nodes (none of which is a singleton). See that no
subgames exist for PD but the whole game.
,
j
I
I
I
I~~~~.
I \i
~~:"....,..
=--~------.•
1 •••••
c
Fig. 22.9(2): Information Sets and Subgames
In the figure above each player A, B, C has three moves: {V, M, D} ; {u, rn,
d} ; {D, Jl, ()}.
• At the first stage A has one information set (singleton).
• At the second stage B knows that A either chooses {D} or {V,M} that are
the two information sets of B.
• At the third stage C knows that B has chosen among {Dd,Dm} , {Du},
{Md,Mm,Vd,Vm}, [Mu.Uu} that are four C's information sets.
Therefore, the game has three subgames (two proper subgames): i) the
original game, ii) a subgame following D, iii) a subgame following the path
D-u.
Through the concept of information set game theory can face strategic
interactions that are partly sequential and partly simultaneous moves.
94
r
1.0
2,3
By definition
• A has four strategies: {UV, UE, DV, DE}
• B has 'two strategies: {u, d}
Although nv
and DE are sequences of moves never played by A, they
cannot be discarded.
Convertion of Extensive Games to Normal Form
B
u cl
1.0
uv 4,1 1,0
2,3
96
Dynamic Games with
Complete
Information
With three players we represent the game in normal form. See that the
representation of the game is done with two matrices as given below:
A plays U A plays D
C C
Left Right Left Right
u 3, 1, 1 3, 1, 1 B u 1,2,2 2,0,3
B 0,0,0 4,2,1
d 3, 1, 1 3, 1, 1 d
7 22
.1
dill
C
u/:'-
B/' _1 ri8~ 0,3 left right
·C;,
\7" 010
B
u 2,2 013
d O,ft 1,1
2,1
It can be seen that "right" is a dominant strategy for player C and the matrix of
the subgame has the only NE {d,right}. That excludes {U,u,left} from being a
SPNE. The only SPNE of the original game is {D,d,right} leading to the
. payoff (4,2,1).
.........................................................................................
Player 2 Player2
L2 M2 L2 M
2
Player 1 Ll (1, 1) (5,0) Player 1 Ll (2,2) (6, 1)
Ml (0,5) (4,4) Ml (1,6) (5,5)
Suppose the players play the game twice, and they observe the outcome of the
first game before playing it once again. We further assume that the payoff of
the entire game, taking two stages together is simply the sum of the payoffs
from the two stages (we assume there is no discounting, that is, the same
amount of payoff in any period gives the same amount of utility. Generally, if
we think in monetary terms we will prefer Rs. 100 today than Rs. 100
tomorrow. This is because there is a discount rate, which is embedded in our
inter-temporal choice decisions).
It is easy to see that the Nash equilibrium of the first stage of the game is (Ll,
L2) with payoff (1, I); we have added the equilibrium payoff of the first stage
to the payoffs of the game in the next stage. In the second stage also there is a
unique N ash equilibrium of the two-stage prisoners' dilemma.
Defmition: Given a stage game G, let G(T) denote the finitely repeated game
in which G is played T times, with the outcome of all the previous games
observed before the next play begins. The payoff for G(T) is simply the sum
of the payoffs from the T stage game.
Proposition: If a stage game G has a unique Nash equilibrium then, for any
finite T, the repeated game G(T) has a unique sub-game perfect Nash
equilibrium: the Nash equilibrium of G being played in every stage of the
game.
99
Non-Cooperative
Player 2
Game Theory-I
L2 M2 R2
Player Ll (1,1) (5,0) (0,0)
1
MI (0,5) (4,4) (0,0)
RI (0,0) (0,0) (3,3)
As a result of adding the two strategies and the distribution of payoffs now
there are two pure strategy Nash equilibrium, namely, (LI, L2) and (RI, R2).
Suppose the above stage game is played twice with the first stage outcome
being observed before the second stage begins.
Since the stage game has more than one Nash equilibrium, it is now possible
for the players to anticipate that different first stage outcomes will be followed
by different stage game equilibria in the second stage. Let us suppose for
example the players anticipate that (RI, R2) will be the second stage outcome
if the first stage outcome is (MI, M2), but (LI, L2) will be the second stage
outcome if anyone of the eight other first stage outcomes occur. Thus, the
game reduces to the following one shot game, where (3, 3) has been added to
the (MI, M2) cell and (1, 1) has been added to rest of the cells.
Player 2
L2 M2 R2
LI (2,2) (6,1) (1,1)
Player 1
MI (1,6) (7,7) (1,1)
RI (1,1) (1,1) (4,4)
There are three pure strategy Nash equilibrium of the above game (LI, L2),
(MI, M2) and (RI, R2). Let us denote the outcome of the repeated game as
[(w, x): (y, z)]; where (w, x) is the first stage outcome and (y, z) is the second
stage outcome. Therefore, the Nash equilibria namely, (LI, L2), (MI, M2)
and (RI, R2) can be achieved in the simplified one shot game if the outcomes
of the repeated game are [(LI, L2), (LI, L2)], [(MI, M2); (RI, R2)] and [(RI,
R2), (LI, L2)] respectively (if the first stage outcome is (LI, L2), the second
stage outcome has to be (LI, L2) according to the players' anticipation and so
on for each of the one shot game's Nash equilibria). The first and the last
Nash equilibria of the repeated game are sub-game perfect and simply
concatenate the Nash equilibrium outcome of the stage game. But the Nash
equilibrium (MI, M2) of the one shot game is possible if the outcome of the
repeated game has the sub-game perfect outcome [(MI, M2); (RI, R2]],
which means in the first stage, the players chose (MI, M2) which is not a
Nash equilibrium of the stage game. We can conclude that cooperation can be
achieved in the first stage of a sub-game perfect outcome of a repeated game.
Proposition: We extend this idea to a stage game being played T times, where
T is any finite number. If G = {AI, A2... , An; UI, U2,..., Un}is a static game of
complete information with multiple Nash equilibria, then there may be sub-
game perfect outcomes of the repeated game G(T) in which for any t< T, the
outcome in stage t is not a Nash equilibrium of the stage game G.
The main points to extract from the above example is that credible threats or
promises about the future behaviour can affect current behavior sub-game
perfection (as we described in the previous unit) may not be a definition
100 strong enough to embody credibility.
22.5.3 Infinitely Repeated Games Dynamic Games with
Complete
In the finite horizon case, the main focus was that credible threats or promises Information
about future behaviour can influence current behaviour and if there are
multiple Nash equilibria of the stage game, then there may be sub-game
perfect outcomes of the repeated game G (T) in which, for any t < T, the
outcome of a stage game is not a Nash equilibrium of G. Whereas in case of
infmitely repeated games a stronger result is true. Even if the stage game has a
unique Nash equilibrium, there could be sub-game perfect outcome of the
infmitely repeated game in which no stage's outcome is a Nash equilibrium of
the stage game G.
Player 2
L2 M2
101
Non-Cooperative In the ehstage if the outcome of all the preceding (t-l) has been (Ml, M2) then
Game Theory-I play Mi in the tthstage otherwise play u. (i = 1, 2) (This type of strategy is
called a trigger strategy). If both the player follows this trigger strategy, then
the outcome of the infinitely repeated game would be (Ml, M2) in every stage
of the game. Now we will prove that given some conditions on the value of 8,
the above trigger strategy is a N ash equilibrium of the infinitely repeated
game, and such an equilibrium is sub-game-perfect.
To show that the above described trigger strategy is a Nash equilibrium of the
game, we have to show that if the player i adopts the trigger strategy, the best
response of the player j is to adopt the same strategy. The present value of the
payoffs that will be generated, for the j'hplayer, if both of the players stick to
the trigger strategy is given by
234
V =4+48+48 +48 ...=--.
1-8
If the fhplayer deviates from the trigger strategy, that is, she plays Lj in the
"
first stage, which will eventually lead to non-cooperation from player i in the
second stage (Li) and consequently from player j (Lj) also, the discounted
payoff of the fh player is given by (the payoff of the first stage for the lh
th
player would be 5, as the i player in the first period following her trigger
strategy will play Ri, and for the remaining periods the payoff of the player t
2
would be 1)V' = 5 + la + la + la ••• = 5 + ~a
3
1-
Therefore, playing Mj, or following the trigger strategy is optimal for the lh
player given that the ith player .sticks to her trigger strategy if and only if
_4_ ~
i -a
5 + _a_ ~
r-a
a ~~.4
Therefore, the trigger strategy 0 is the N ash
equilibrium for both the players if and only if a ~ ~.
4
• subgames in which all the outcomes of the earlier stages have been (RI,
R2)
• sub-games in which the outcome of at least one stage differs from (RI,
R2).
If the players adopt the trigger strategy for the game as a whole, then (i) the
players' strategies in a sub-game in the fist case are again the trigger strategy,
which we have already shown to be the Nash equilibrium of the game as a
whole.
The players' strategies in a sub-game in the second case are simply to repeat
the stage game equilibrium (Lj, L2) forever, which is also Nash equilibrium of
the game as a whole. Therefore, the trigger strategy Nash equilibrium of the
infmitely repeated prisoners' dilemma is sub-game-perfect.
The profit to one firm when both the firms produce (qn/2) is denoted byrr;
i.e. rrm = qm Ca _ q . _ c) = (a-c)2
, 2' 2 m 8
Finally, if firm 1 is going to produce (qu/2) this period, then the quantity that
maximises j's profit in this period, is obtained solving the following simple
. qm
maximisation problem: maxqj (a-~-c) qj. The solution of the problem can
be obtained from the first order condition of profit maximisation, that is,
(a - 2q , ..!. q)
...!!!. - C = O. So, qJ' = a---c
2
qm (a-c)
Or q. = a---c2.2 3(a-c) with
:::;;:-_.
J 2 . 2 'J 2 8
9
associated profIt (a-c)2. We will denote this profit by rrd(d stands for
, 64
deviation).
Therefore, it is a Nash equilibrium for both the firms to play the trigger
strategy, given earlier provided that,present value of payoff from the trigger
strategy 2: present value of the payoffs deviated in the first period. .
1 1 m 8
Or -'-rr
'1-8 2
>
-
rrd +-n
1-8 c
Substituting the values of !id' and z; into the above equation, we get if
6 2: :7'
then the inequality will hold and the tri~ger strategy will be sub-game
perfect.
Thus, we see that collusion in infinitely repeated games can fetch extra
payoffs to the firms.
104
Dynamic Games with
Check Your Progress 4 . Complete
1) What is a strategy in a repeated game? Information
........................................................................................
........................................................................................
........................................................................................
........................................................................................
........................................................................................
2) What do mean by a stage game?
.........................................................................................
........................................................................................
........................................................................................
.........................................................................................
.......................... ' .
3) An infinitely repeated game promising A per period with continuing
cooperation, B for the one period gain from cheating, and C per period
for the non-cooperative solution (B > A> C) will have a subgame perfect
Nash equilibrium in trigger strategies providing-----
.........................................................................................
........................................................................................
........................................................................................
........................................................................................
........................................................................................
4) A cartel-like collusive solution can be a Nash equilibrium only in price-
setting games with -------
........................................................................................
........................................................................................
........................................................................................
.........................................................................................
5) a) What is the equilibrium of the following game if it is played as a
I
Player A
1
,
.,
i'
8
"",
Down
" -, -, -, 3 " -,
-, -,
5
4
6
" -, -,
-,
105
Non-Cooperative
Game Theory-I
The unit discussed repeated games that analyses whether threats and promises
about future can affect the current behavior. Besides learning the presentation
of these games, we have been exposed to their solution mechanismsby
considering finite and infinite time horizons. In the process we have seen the
change that occurs in one shot games like prisoner's dilemma. The use
oftrigger strategy where a player initially cooperates but punishes the
opponent if observed a certain level of defection. Bilateral bargaining is seen
as an attempt to resolve a bargaining situation, i.e., to determine the particular
form of cooperation and the corresponding payoffs for two players involved.
Using the framework of infinitely repeated games we have seen the process of
collusion fetching extra payoff.
106
Dynamic Games with
22.7 KEY WORDS Complete
Information
Complete Information: Knowledge about other players is available to. all
participants.
Incredible Threat: A threat made by a player in a Sequential Game which
would not be in the best interest for the player to carry out.
Backward Induction: Backward induction is a technique to solve a game of
perfect information. It first considersthe moves that are the last in the game,
and determines the best move for the playerin each case.
Common Knowledge: A fact which all players know.All. know that they all
know it, andso on.
Dominating Strategy: A strategy dominates another strategy of a player if it
always gives a better payoff tothat player, regardless of what the other players
are doing. .
Extensive Game: A game that describes with a tree how a game is
played. Game: A game is a formal description of a strategic situation.
Mixed Strategy: An active randomization, with given probabilities, .that
determines theplayer's decision.
Perfect Information: At any point in time in the game only one player makes
a move,and knows all the actions that have been made until then.
Rationality: A player seeks to play in a manner which maximizes her
ownpayoff. .
Subgame Perfect Equilibrium; An equilibrium in which all actions are Nash
equilibria for all subgames.
Perfect Information: Each information set consists of onlyone node.
Perfect Recall: Each player recalls exactly what he did in thepast.
Discount Factor: If the rate of return is r (which is generally positive), then
1
the ratio- - is called the discount factor. The discount factor in game theory
l+r
literature is generally denoted by a.
Finitely Repeated Games: When a game is repeated fmite number of times
.
it is called a finitely repeated game .
Infinitely Repeated Games: When a game is repeated infinite number of
times it is called a fmitely repeated game.
Present Value: A future asset evaluated at present time is called the present
.value of the asset. If t periods later an asset is worth A, then it's present value
now is given by the formula- A )t.
C l+r
Stage Game: In a repeated game the initial game, which is repeated in every
time period is called the stage game.
Stackelberg Competition: A strategic game in which the leader firm moves
first and then the follower firms move sequentially.
22.12 EXERCISES
1) The simultaneous move game is played twice with the outcome of the
first stage being observed before playing the second stage. There is no
discounting factor. Can the payoff (4,4) be achieved in the first stage in
a pure strategy sub-game perfect Nash equilibrium? If so, give strategies
that do so. If not then, why not?
Player 2
Left Center Right
Top 3,1 0,0 5,0
Middle 2,1 1,2 3,1
Player 1 Bottom 1,2 0,1 4,4
Thus, the trigger strategy is a Nash equilibrium, but to show that it is sub
game perfect Nash equilibrium, you need to show that the trigger strategy
induces Nash equilibrium in every sub-game of the infinitely repeated
Bertrand duopoly.
There are two possible situations
Case I: No one deviates at any point. Then the sub-game will look: like the
game itself, (pm, pm) is chosen in all earlier stages and the strategy profile
induces (pm, pm) in the subsequent games, which we have shown to be
Nash equilibrium provided the restriction on a.
Case II: If something else happens in the game, then the strategy will
induce (c, c) at all subsequent games, which we have shown to be a Nash
equilibrium ill the previous unit.
111