Solution Concepts PDF
Solution Concepts PDF
Jonathan Levin
April 2006
These notes discuss some of the central solution concepts for normal-
form games: Nash and correlated equilibrium, iterated deletion of strictly
dominated strategies, rationalizability, and self-confirming equilibrium.
1 Nash Equilibrium
Nash equilibrium captures the idea that players ought to do as well as they
can given the strategies chosen by the other players.
C D
C 1, 1 −1, 2
D 2, −1 0, 0
B F
B 2, 1 0, 0
F 0, 0 1, 2
There are two pure Nash equilibria (B, B) and (F, F ) and a mixed
strategy equilibrium where Row plays 23 B + 13 F and Column plays
1 2
3B + 3F.
1
3. Payoff functions: for each i = 1, ..., I, ui : S → R
ui (σ i , σ −i ) ≥ ui (si , σ −i ).
2 Correlated Equilibrium
2.1 Equilibria as a Self-Enforcing Agreements
Let’s start with the account of Nash equilibrium as a self-enforcing agree-
ment. Consider Battle of the Sexes (BOS). Here, it’s easy to imagine the
players jointly deciding to attend the Ballet, then playing (B, B) since nei-
ther wants to unilaterally head off to Football. However, a little imagination
suggests that Nash equilibrium might not allow the players sufficient freedom
to communicate.
Example 2, cont. Suppose in BOS, the players flip a coin and go to the
Ballet if the coin is Heads, the Football game if Tails. That is, they
just randomize between two different Nash equilibria. This coin flip
allows a payoff ( 32 , 32 ) that is not a Nash equilibrium payoff.
2
So at the very least, one might want to allow for randomizations between
Nash equilibria under the self-enforcing agreement account of play. More-
over, the coin flip is only a primitive way to communicate prior to play. A
more general form of communication is to find a mediator who can perform
clever randomizations, as in the next example.
Example 3 This game has three Nash equilibria (U, L), (D, R) and ( 12 U +
1 1 1 5 5
2 D, 2 L + 2 R) with payoffs (5, 1), (1, 5) and ( 2 , 2 ).
L R
U 5, 1 0, 0
D 4, 4 1, 5
Thus the players will follow the mediator’s suggestion. With the me-
diator in place, expected payoffs are ( 10 10
3 , 3 ), strictly higher than the
players could get by randomizing between Nash equilibria.
3
• A finite set of states Ω
• A probability distribution p on Ω.
Example 2, cont. In the BOS example with the coin flip, the states are
Ω = {Heads, Tails}, the probability measure is uniform on Ω, and Row
and Column have the same partition, {{Heads}, {Tails}}.
Example 3, cont. In this example, the set of states is Ω = {1, 2, 3}, the
probability measure is again uniform on Ω, Row’s partition is {{1}, {2, 3}},
and Column’s partition is {{1, 2}, {3}}.
This definition requires that fi maximize i’s ex ante payoff. That is, it
treats the strategy as a contingent plan to be implemented after learning the
partition element. Note that this is equivalent to fi maximizing i’s interim
payoff for each Hi that occurs with positive probability – that is, for all
i,ω, and every s0i ∈ Si ,
X X
ui (fi (ω), f−i (ω 0 ))p(ω 0 |hi (ω)) ≥ ui (s0i , f−i (ω 0 ))p(ω 0 |hi (ω))
ω0 ∈hi (ω) ω 0 ∈hi (ω)
Here, p(ω 0 |hi (ω)) is the conditional probability on ω 0 given that the true
state is in hi (ω). By Bayes’ Rule,
4
The definition of CE corresponds to the mediator story, but it’s not very
convenient. To search for all the correlated equilibria, one needs to con-
sider millions of mechanisms. Fortunately, it turns out that we can focus
on a special kind of correlating mechanism, callled a direct mechanism. We
will show that for any correlated equilibrium arising from some correlating
mechanism, there is a correlated equilibrium arising from the direct mech-
anism that is precisely equivalent in terms of behavioral outcomes. Thus
by focusing on one special class of mechanism, we can capture all possible
correlated equilibria.
The result is trivial if there is only one information set Hi in the original
mechanism for which fi (Hi ) = si . In this case, conditioning on si is the same
as conditioning on Hi in the original. More generally, we substitute for q to
obtain: X
1
· ui (s0i , f−i (ω))p(ω).
Pr(fi (ω) = si )
ω|fi (ω)=si
5
Thus what really matters in correlated equilibrium is the probability
distribution over strategy profiles. We refer to any probability distribution
q over strategy profiles that arises as the result of a correlated equilibrium
as a correlated equilibrium distribution (c.e.d.).
Example 2, cont. In the BOS example, the c.e.d. is 12 (B, B), 12 (F, F ).
Example 3, cont. In this example, the c.e.d is 13 (U, L), 13 (D, L), 13 (D, R).
The next result characterizes correlated equilibrium distributions.
Proposition 3 The distribution q ∈ ∆(S) is a correlated equilibrium dis-
tribution if and only if for all i, every si with q(si ) > 0 and every s0i ∈ Si ,
X X
ui (si , s−i ) q (s−i |si ) ≥ ui (s0i , s−i )q(s−i |si ). (2)
s−i ∈S−i s−i ∈S−i
Proof. (⇐) Suppose q satisfies (2). Then the “obedient” profile f with
fi (s) = si is a correlated equilibrium given the direct mechanism (S, {Hi }, q)
since (2) says precisely that with this mechanism si is optimal for i given
recommendation si . (⇒) Conversely, if q arises from a correlated equilib-
rium, the previous result says that the obedient profile must be a correlated
equilibrium relative to the direct mechanism (S, {Hi }, q). Thus for all i
and all recommendations si occuring with positive probability, si must be
optimal – i.e. (2) must hold. Q.E.D.
6
Property 2 Correlated equilibria exist in finite games.
L R
U 5, 1 0, 0
D 4, 4 1, 5
7
2.4 Comments
1. The difference between mixed strategy Nash equilibria and correlated
equilibria is that mixing is independent in NE. With more than two
players, it may be important in CE that one player believes others
are correlating their strategies. Consider the following example from
Aumann (1987) with three players: Row, Column and Matrix.
0, 0, 3 0, 0, 0 2, 2, 2 0, 0, 0 0, 0, 0 0, 0, 0
1, 0, 0 0, 0, 0 0, 0, 0 2, 2, 2 0, 1, 0 0, 0, 3
No NE gives any player more than 1, but there is a CE that gives
everyone 2. Matrix picks middle, and Row and Column pick (Up,Left)
and (Down,Right) each with probability 12 . The key here is that Matrix
must expect Row to pick Up precisely when Column picks Left.
2. Note that in CE, however, each agent uses a pure strategy – he just
is uncertain about others’ strategies. So this seems a bit different than
mixed NE if one views a mixed strategy as an explicit randomization
in behavior by each agent i. However, another view of mixed NE if
that it’s not i’s actual choice that matters, but j’s beliefs about i’s
choice. On this account, we view σ i as what others expect of i, and i
as simply doing some (pure strategy) best response to σ −i . This view,
which is consistent with CE, was developed by Harsanyi (1973), who
introduced small privately observed payoff perturbations so that in
pure strategy BNE, players would be uncertain about others behavior.
His “purification theorem” showed that these pure strategy BNE are
observably equivalent to mixed NE of the unperturbed game if the
perturbations are small and independent.
3. Returning to our pre-play communication account, one might ask if a
mediator is actually needed, or if the players could just communicate
by flipping coins and talking. With two players, it should be clear
from the example above that the mediator is crucial in allowing for
messages that are not common knowledge. However, Barany (1992)
shows that if I ≥ 4, then any correlated equilibrium payoff (with ratio-
nal numbers) can be achieved as the Nash equilibrium of an extended
game where prior to play the players communicate through cheap talk.
Girardi (2001) shows the same can be done as a sequential equilibrium
provided I ≥ 5. For the case of two players, Aumann and Hart (2003)
characterize the set of attainable payoffs if players can communicate
freely, but without a mediator, prior to playing the game.
8
3 Rationalizability and Iterated Dominance
Bernheim (1984) and Pearce (1984) investigated the question of whether one
should expect rational players to introspect their way to Nash equilibrium
play. They argued that even if rationality was common knowledge, this
should not generally be expected. Their account takes a view of strategic
behavior that is deeply rooted in single-agent decision theory.
To discuss these ideas, it’s useful to explicitly define rationality.
2. Beliefs cannot conflict with other players being rational, and being
aware of each other’s rationality, and so on (but they need not be
correct).
b1 b2 b3 b4
a1 0, 7 2, 5 7, 0 0, 1
a2 5, 2 3, 3 5, 2 0, 1
a3 7, 0 2, 5 0, 7 0, 1
a4 0, 0 0, −2 0, 0 10, −1
9
• Row will play a3 if Column plays b1
• Column will play b1 if Row plays a1
• Let Si0 = Si
½ ¾
k si ∈ Sik−1 : There is no σ i ∈ ∆(Sik−1 ) s.t.
• Let Si = k−1
ui (σ i , s−i ) > ui (si , s−i ) for all s−i ∈ S−i
T
• Let Si∞ = ∞ k=1 Si .
k
10
Iterated strict dominance never eliminates Nash equilibrium strategies,
or any strategy played with positive probability in a correlated equilibrium
(proof left as an exercise!). Indeed it is often quite weak. Most games,
including many games with a unique Nash equilibrium, are not dominance
solvable.
Example 4, cont. In this example, b4 is strictly dominated. Eliminating
b4 means that a4 is also strictly dominated. But no other strategy can
be eliminated.
Proposition 5 In finite games, iterated strict dominance and correlated
rationalizability give the same solution set, i.e. Si∞ = Ri .
This result is suggested by the following Lemma (proved in the last
section of these notes).
Lemma 1 A pure strategy in a finite game is a best response to some beliefs
about opponent play if and only if it is not strictly dominated.
Proof of Proposition.
R ⊂ S ∞ . If si ∈ Ri , then si is a best response to some belief over R−i .
Since R−i ⊂ S−i , Lemma 1 implies that si is not strictly dominated. Thus
Ri ⊂ Si2 for all i. Iterating this argument implies that Ri ⊂ Sik for all i, k,
so Ri ⊂ Si∞ .
S ∞ ⊂ R. It suffices to show that S ∞ is a best-reply set. By definition,
no strategy in S ∞ is strictly dominated in the game in which the set of
actions is S ∞ . Thus, any si ∈ Si∞ must be a best response to some beliefs
∞.
over S−i Q.E.D.
3.3 Comments
1. Bernheim (1984) and Pearce (1984) originally defined rationalizability
assuming that players would expect opponents to mix independently.
So B is a best reply set if ∀si ∈ Bi , there is some σ −i ∈ ×j6=i ∆(Sj ) to
which si is a best reply. For I = 2, this makes no difference, but when
I ≥ 3, their concept refines ISD (it rules out more strategies).
2. Brandenburger and Dekel (1987) relate correlated rationalizability to
subjective correlated equilibrium. While SCE is more permissive than
rationalizability, requiring players to have well defined conditional be-
liefs (and maximize accordingly) even for states ω ∈ Ω to which they
assign zero probability leads to a refinement of SCE that is the same
as correlated rationalizability.
11
3. One way to categorize the different solution concepts is to note that if
one starts with rationality, and common knowledge of rationality, the
concepts differ precisely in how they further restrict the beliefs of the
players about the distribution of play. Doug Bernheim suggests the
following table:
12
which can be related to the following “dual” problem:
m
X
max
m
yi bi (7)
y∈R+
i=1
n
à m
!
X X
s.t. cj − yi aij =0 ∀j = 1, ..., n.
j=1 i=1
Theorem 1 Suppose problems (3) and (7) are feasible (i.e. have non-empty
constraint sets). Then their solutions are the same.
Problem I:
min δ
σ −i ,δ
s.t. σ −i (s−i ) ≥ 0 ∀s−i ∈ S−i
X X
σ −i (s−i ) ≥ 1 and − σ −i (s−i ) ≥ −1
s−i s−i
X £ ¤
δ+ σ −i (s−i ) ui (si , s−i ) − ui (s0i , s−i ) ≥ 0 ∀s0i ∈ Si
s−i
Observe that si is a best response to some beliefs over opponent play if any
only if the solution to this problem is less than or equal to zero.
Problem II:
max ε1 − ε2
η,ε1 ,ε2 ,σi
|S | |S |
s.t. ε1 , ε2 ∈ R+ , σ i ∈ R+ i , η ∈ R+ −i
X
σ i (s0i ) ≥ 1
s0i
X £ ¤
η(s−i ) + ε1 − ε2 + σ i (s0i ) ui (si , s−i ) − ui (s0i , s−i ) = 0 ∀s−i ∈ S−i
s0i
13
Observe that si is strictly dominated if and only if the solution to this
problem is strictly greater than zero – i.e. si is not strictly dominated if
and only if the solution to this problem is less than or equal to zero.
Finally, the Duality Theorem for linear programming says that so long
as these two problems are feasible (have non-empty constraint sets), their
solutions must be the same, establishing the result. Q.E.D.
4 Self-Confirming Equilibria
The third possible foundation for equilibrium is learning. We’ll look at
explicit learning processes later; for now, we ask what might happen as the
end result of a learning processes. For instance, if a learning process settles
down into steady-state play, will this be a Nash Equilibrium? Fudenberg
and Levine (1993) suggest that a natural end-result of learning is what they
call self-confirming equilibria. In a self-confirming equilibrium:
2. Beliefs cannot conflict with the empirical evidence (i.e. must match
the empirical distribution of play).
14
In extensive form games, the situation is different, as the next example
shows.
Example 2 In the entry game below, the only Nash equilibria are (Enter,
Don’t Fight) and (Don’t Enter, Fight). These equilibria, with correct
beliefs, are also self-confirming equilibria.
Player 1
-1, 1 1, 1
An Entry Game
There is also another SCE, however, where Player 1 plays Don’t Enter
and believes Fight, while Player 2 player plays Don’t Fight and be-
lieves Don’t Enter. In this SCE, player 1 has the wrong beliefs, but
since he never enters, they’re never contradicted by the data!
In the entry game, the non-Nash SCE is indistinguishable from a Nash
equilibrium in terms of observed behavior. But even that need not be the
case, as the next example shows.
Example 3 (Fudenberg-Kreps, 1993) Consider the three player game be-
low.
1 A1 2 A2
(1,1,1)
D1 D2
3
L R L R
15
In this game, there is a self-confirming equilibrium where (A1 , A2 ) is
played. In this equilibrium, player 1 expects player 3 to play R, while
player 2 expects 3 to play L. Given these beliefs, the optimal strategy
for player 1 is to play A1 , while the optimal strategy or player 2 is to
play A2 . Player 3’s beliefs and strategy can be arbitrary so long as the
strategy is optimal given beliefs.
The key point here is that there is no Nash equilibrium where (A1 , A2 )
is played. The reason is that in any Nash equilibrium, players 1 and
2 must have the same (correct) beliefs about player 3’s strategy. But
if they have the same beliefs, then at least one of them must want to
play D.
Example 4 Consider the following two-player game. Suppose that after the
game is played, the column player observes the row player’s action, but
the row player observes only whether or not the column player chose
R, and gets no information about her payoff.
L M R
U 2, 0 0, 2 0, 0
D 0, 0 2, 0 3, 3
This game has a unique Nash equilibrium, (D, R); indeed the game is
dominance-solvable. The profile (D, R) is self-confirming too; however
that is not the only self-confirming profile. The profile (U, M ) is also
self-confirming. In this SCE, column has correct beliefs, but row be-
lieves that column will play L. This mistaken belief isn’t refuted by
the evidence because all row observes is that column does not play R.
There are also SCE where row mixes, and where both players mix.
Note that in the (U, M ) SCE, row’s beliefs do not respect column’s ra-
tionality. This suggests that one might refine SCE by further restrict-
ing beliefs to respect rationality or common knowledge of rationality
– Dekel, Fudenberg and Levine (1998) and Esponda (2006) explore
this possibility.
16
The above example is somewhat contrived, but the idea that players
might get only partial feedback, and this might affect the outcome of learn-
ing, is natural. For instance, in sealed-bid auctions it is relatively common
to announce only the winner and possibility not even the winning price, so
the information available to form beliefs is rather limited.
The way to read this is that for each information set at which some player
j 6= i moves, player i’s belief puts a point mass on the probability distrib-
ution that exactly coincides with the distribution induced by j’s strategy.
Thus i has correct beliefs at all opponent information sets.
17
References
[1] Aumann, R. (1974) “Subjectivity and Correlation in Randomized
Strategies,” J. Math. Econ.
18