Combined Lecture Notes
Combined Lecture Notes
MARTIN CRIPPS
The purpose of this topic is to teach you what a game in strategic form is and how
economists analyze them.
Example 2: When firms choose quantities Si = [0, ∞) is the set of possible quantities. A
random choice of a quantity can be represented by a probability distribution over the set of
positive numbers. One way of describing such a distribution is to write down its cumulative
distribution function (cdf) F (x) := P r(Firm’s outputs less than or equal to x).
Payoffs from Mixed Actions: A player’s payoff when mixed actions are played is an expec-
tation (or average) taken over all the payoffs they may get multiplied by the probability
they get them. This expectation is taken assuming the players randomise independently.
Examples of payoffs from mixed actions (Rock Paper Scissors): Suppose in this game you
are the column player and you believe the row player will play action R with probability
p, P with probability q and S with probability 1 − p − q. You are interested in your payoff
from action P. Below we have written out the payoffs and emphasised the relevant numbers
for the column player.
R P S
p 0,0 -1,1 1,-1
q 1,-1 0,0 -1,1
1 − p − q -1,1 1,-1 0,0
If she plays P she expects to get 1 with probability p, 0 with probability q and −1 with
probability 1 − p − q. Thus on average she expects to get
p × 1 + q × 0 + (1 − p − q) × (−1) = 2p + q − 1.
If we repeat these calculations for each column we can work out what each action will give
her in expectation
R P S
p 0,0 -1,1 1,-1
q 1,-1 0,0 -1,1
1−p−q -1,1 1,-1 0,0
1 − p − 2q 2p + q − 1 q − p
2. Dominance
2.1. Strict Dominance.
Definition 1. A mixed strategy σi strictly dominates the pure action s0i for player i, if and
only if, player i’s payoff when she plays σi and the other players play actions s−i is strictly
higher than her payoff from s0i against s−i for any actions s−i the others may play:
ui (σi , s−i ) > ui (s0i , s−i ), ∀s−i .
Consider the following game (we only put in the row player’s payoffs as that is all that
matters for now).
L R
T 3 0
M 0 3
B 1 1
4 MARTIN CRIPPS
And consider two strategies for the row player: play the first two rows with equal probability
σi = (1/2, 1/2, 0), or play the bottom row si = B. We can write the expected payoffs to
these strategies in the following way.
L R L R
1
2 3 0 0 3 0
1
2 0 3 0 0 3
0 1 1 1 1 1
3 3 1 1
2 2
From this you can see that if the column player plays L the strategy σi gives the row player
the payoff of 32 but the strategy si gives the row player the payoff 1. And, if the column
player plays R strategy σi also gives the row player the payoff of 23 but the strategy si gives
the row player the payoff 1. Thus the strategy σi always does better than the strategy
si . To describe this we say si is strictly dominated and we would never expect a rational
player to play this.
However, eliminating strictly dominated actions can allow us to make strong predictions
about what actions the players will use. Consider the following game. . .
L R
U (8,10) (-100,9)
D (7,6) (6,5)
First observe that R is strictly dominated by L for the column player. So a rational column
player will never play R. If the row player knows this then they should play U getting
8 rather than 7. So we predict (U, L) as the outcome of this game. But, some types
of players may be very worried about the −100. If you had some doubts about column
player’s rationality would you be willing to play U ?
2.2. Weak Dominance. The notion of weak domination does not require players to
strictly prefer one strategy to another, it is enough to weakly prefer one to the other.
Definition 2. A mixed strategy σi weakly dominates the pure action si ∈ Si for player i,
if and only if, playing σi is at least as good as playing si whatever the other players do:
ui (σi , s−i ) ≥ ui (si , s−i ), ∀s−i .
L R
T (1,1) (0,0)
M (0,0) (0,0)
B (0,0) (0,0)
T weakly dominates M for the row player, because T is better than M if column plays L
and T is no worse than M is column plays R.
1: GAMES IN STRATEGIC FORM 5
L M R
T (3,3) (2,2) (5,2)
M (2,10) (4,4) (6,3)
B (2,1) (2,7) (1,0)
• Step 1: R is strictly dominated by L, so a rational column player will not play R
L M R
T (3,3) (2,2) -
M (2,10) (4,4) -
B (2,1) (2,7) -
• Step 2: In the remaining game 12 T + 12 M strictly dominates B. So if the row player
knows the column player is rational and is rational themselves, then they will not
play B.
L M R
T (3,3) (2,2) -
M (2,10) (4,4) -
B - - -
• Step 3: In the remaining game L strictly dominates M for the column player. So if
the column player is rational, knows that the row player is rational and knows that
the row player knows the column player is rational, then they will not play M.
L M R
T (3,3) - -
M (2,10) - -
B - - -
• Step 4: In the remaining game T strictly dominates M for the row player. So if the
row player is rational and the row knows that the column player is rational, and
the row knows that the column player knows the row player is rational, then they
will not play M.
L M R
T (3,3) - -
M (2,10) - -
B - - -
Summary: the set of strategies that survive the iterated elimination of strictly dominated
actions are those actions that can be played by rational players with common knowledge
of rationality
6 MARTIN CRIPPS
3. Nash Equilibrium
Definition 3. A action profile s∗ = (s∗1 , . . . , s∗I ) is a Nash Equilibrium if s∗i is a best
response to the actions (s∗−i ) of the other players, and this is true for all players i. A
mixed strategy profile σ ∗ is a Nash Equilibrium if for all players i, if the same condition
holds for mixed strategies.
This game has a mixed strategy Nash equilibrium where each player plays every action with
probability 13 . If one player plays each action with equal probability, then their opponent
gets zero from any action. Every action they play gives them zero,, so every action is a
best response to this strategy. One best response is to play ( 31 , 13 , 13 ).
Example 2 : (Matching Pennies)
H T
H (-1,1) (1,-1)
T (1,-1) (-1,1)
If the column player plays ( 12 , 12 ) then the row player expects to get zero from H and from
T . Thus, any action is optimal. It follows that ( 21 , 12 ) is a best response by the row player.
(This is a general property of mixed equilibria—one player’s actions at the equilibrium
make their opponent indifferent and therefore willing to randomize.)
You might think that this example is silly because the players ought to be able to avoid this
Nash equilibrium. But suppose we increase the number of players so there are three players
(player 3 chooses the matrix) then there still is a weakly dominated Nash equilibrium.
3 Plays L L R 3 Plays R L R
U (1,1,1) (0,0,0) U (0,0,0) (0,0,0)
D (0,0,0) (0,0,0) D (0,0,0) (0,0,0)
This Nash equilibrium seems much harder to avoid as one of the player (say player 3) now
has to persuade both of the others to change there actions. There are, therefore, sensible
situations where players end up playing weakly dominated actions at a Nash equilibrium.
4.2. Method 2 : The underlining method. This is a process for finding all the pure-
strategy Nash equilibria of a game. You only use it in finite games. Here is an example of
8 MARTIN CRIPPS
4.3. Method 3 : Mixed strategy NE’s in 2 × 2 games. This is a process that will
help you find Nash equilibria in games with 2 players each of which has 2 actions. To find
mixed strategy Nash equilibria is difficult and will usually require a computer for n × n
games. To find a mixed we will consider a specific game below, which is often called the
Battle of the Sexes
L R
T (3,1) (0,0)
B (0,0) (1,3)
Here is the recipe you follow:
• Step 1: Assume that the column player plays a mixed strategy (q, 1−q)and evaluate
the row player’s payoffs to all their actions.
q 1−q
T (3,1) (0,0) = 3q + 0(1 − q)
B (0,0) (1,3) = 0q + 1(1 − q)
1: GAMES IN STRATEGIC FORM 9
A B
A (1,1) (0,0)
B (0,0) (1,1)
These games generally have huge numbers of equilibria (it has two pure strategy equilibria
and one mixed strategy equilibrium). Sometimes we need a tool for selecting among Nash
equilibria in order to make predictions. In this game it seems virtually impossible, any
justification for the (A, A) equilibrium seems like it ought to be a justification for (B, B)
also. However, in some games we can make choices.
J-J. Rousseau introduced the Stag-Hunt game. He had in mind the following story. Two
hunters can either independently hunt a rabbit (this requires no cooperation), or can try
to coordinate to catch a stag. But hunting a stag fails unless we both commit to hunting
a stag. Here are the usual payoffs given to this game. Another important feature is that
hunting a rabbit gets easier if the other player isn’t also hunting a rabbit.
Stag Rabbit
Stag (9,9) (0,8)
Rabbit (8,0) (7,7)
This game has two Nash equilibria: (S, S) and (R, R). The (S, S) is Pareto efficient, but
quite risky because for me to get 9 I need to be quite sure my opponent is also playing S.
The (R, R) NE is safer, because whatever my opponent does I do OK (I get at least 7),
but not Pareto Efficient. We say that: (S, S) is Pareto dominant NE and (R, R) is a Risk
Dominant NE.
Most games have many Nash equilibria. Some people view this a problem and we have to
find a way of choosing a unique way of predicting how the game will be played. Sometimes
the names or history of the strategy mean that particular action is most prominent and
therefore sensible to coordinate on. This is called a Focal Point. For example, at the end of
Slumdog Millionaire the two characters have no mobile phones and must meet somewhere
1: GAMES IN STRATEGIC FORM 11
MARTIN CRIPPS
Here are examples of directed graphs that are not trees. the first violates the unique
A. Perfect Information Games in Extensive Form
path condition and the second violates the unique initial node condition.
Not a tree
A. Perfect Information Games in Extensive Form
Not a tree
The important feature of these graphs (extensive form games) is the way they are la-
belled, because it is this that really explains the game. Here are the rules for labelling:
• Each non-terminal node is labelled with a player’s name and the edges from that
node are actions the player can take at that point in the game.
1
2 MARTIN CRIPPS
• Each terminal node describes a unique history of play. It is labelled with a payoff
to every player in the game for that history.
A. Perfect Information Games in Extensive Form
Here is an example of a tree with such a labelling:
. 1
Fold Raise
2 2
(-1,1) (1,-1)
(0,0)
1
Raise
Fold
(-1,1) (2,-2)
If you wanted to give a verbal description of what goes on in this game. You might proceed
as follows. . . The game begins with Player 1 choosing between two actions Fold or Raise.
If Player 1 chooses Fold, Player 2 observes 1’s action and then must choose between Fold
and Raise and the game ends. The payoffs the players get are either (0, 0) or (−1, 1). If
Player 1 Raises initially, then Player 2 also gets to choose between Fold or Raise. If Player
2 Folds, the game is over and the payoffs are (1, −1). If Player 2 Raises, then Player 1 gets
another go and can choose between Fold or Raise a second time after which the game ends.
If Player 1 folds, the payoffs are (−1, 1) and if Player 1 Raises the payoffs are (2, −2).
I hope you can see A.#Perfect#Information#Games#in#Extensive#Form6
that the picture is much clearer than the words. Also notice the
convention that we describe the payoffs as a vector giving Player 1’s payoff first.
.6 1
Fold'' Raise'
2 2
Raise'
Fold'' Fold'' Raise'
(0,0)'
1
Raise'
Player'1' Player'2' Fold''
We also want to allow for the possibility that there is some randomness in the game.
(Dealing cards, rolling dice, exogenous uncertainty in the model.) To allow for random
2: GAMES IN EXTENSIVE FORM 3
moves we introduce a player called Nature with a name zero. Each edge from a zero
node has a probability attached that describes the probabilities of the random move. For
example, A. Perfect Information Games in Extensive Form
Random Moves 1
Fold Raise
2 2
1/2
1/2
(-1,1) (2,-2)
Important: There is a reason these pictures are called Perfect Information Games in
Extensive Form, this is because there is very little uncertainty in these games. In a perfect
information extensive form game each player knows exactly where they are in the game
when they take a move/decision. (The only thing they don’t know is how future moves
will be played.) They do know exactly what has happened at every point they make must
make a decision. Examples of perfect information games like this are Chess, Backgammon,
Go, Monopoly. However, games such as Poker, Bridge (where the players have private
information), or Paper-Scissors-Rock (where the players move at the same time) cannot be
described in this way—we will get to them a little later.
1.2. Pure Strategies. Thus far we have been describing a situation the players might
find themselves in. Now we want to describe what the players might do.
Definition 1. A Pure Strategy for a player is an instruction book on how to play the
game. This instruction book must be complete and tell the player what to do at every point
at which he/she must make a move. Because we want to consider what happens when
players make mistakes we must even include apparently redundant instructions.
Here is an example of an instruction book (Pure Strategy) for Player 1 in the game we
have been studying
4 A.#Perfect#Information#Games#in#Extensive#Form6
MARTIN CRIPPS
1’s#Instructions6 1
(Fold,Fold)6
Fold'' Raise'
2 2
(11,1)' (1,11)'
(0,0)'
1
Fold'' Raise'
(11,1)' (2,12)'
This instruction book has the instruction Fold at the first decision and Fold at the second
decision. (Clearly, this second instruction is redundant but we still include it.) Here are
some other pure strategies for Player 1 in the game.
A.#Perfect#Information#Games#in#Extensive#Form6
1’s#Instructions6 1
(Raise,Fold)6
Fold'' Raise'
2 2
(11,1)' (1,11)'
(0,0)'
1
Fold'' Raise'
(11,1)' (2,12)'
Player 1 Raises at the first decision node and Folds at the second.
A.#Perfect#Information#Games#in#Extensive#Form6
2: GAMES IN EXTENSIVE FORM 5
1’s#Instructions6 1
(Raise,Raise)6
Fold'' Raise'
2 2
(11,1)' (1,11)'
(0,0)'
1
Fold'' Raise'
(11,1)' (2,12)'
A.#Perfect#Information#Games#in#Extensive#Form6
Player 1 Raises at the first decision node and Raises at the second.
1’s#Instructions6 1
(Fold,Raise)6
Fold'' Raise'
2 2
(11,1)' (1,11)'
(0,0)'
1
Fold'' Raise'
(11,1)' (2,12)'
Player 1 Folds at the first decision node and Raises at the second.
In this game Player 1 has 4 pure strategies: (Raise, Raise), (Raise, Fold), (Fold, Raise),
(Fold, Fold). Notice that Player 2 also has 4 pure strategies, but none of her instructions
are ever redundant.
1.3. Mixed and Behaviour Strategies. From our study of games in strategic form we
also want to allow for the players to make random moves. There are two ways of doing this.
The first is to imagine a library containing all the possible instruction books for Player 1.
(In this case there would be 4 books (Raise, Raise), (Raise, Fold), (Fold, Raise), (Fold,
Fold).) Then choosing one of the books at random, for example Player 1 could choose the
(Raise,Raise) book with probability 21 and the (Fold, Fold) book with probability 12 . This
is an example of what is called a mixed strategy.
Definition 2. A mixed strategy for a player is a random choice of an instruction book
(i.e. a random choice of a pure strategy).
This is a picture of what’s going on
6 MARTIN CRIPPS
A.#Perfect#Information#Games#in#Extensive#Form6 A.#Perfect#Information#Games#in#Extensive#Form6
1’s#Mixed#strategy6 1 1’s#Mixed#strategy6 1
½#(#Raise,#Raise)6 ½#(#Raise,#Raise)6
Fold'' Raise' Fold'' Raise'
½#(#Fold,#Fold)6 ½#(#Fold,#Fold)6
6 6
2 2 2 2
Fold'' Raise' Fold'' Raise' Raise'
Fold'' Fold'' Raise'
Notice that the randomness appears to be occurring only once (at the selection stage)
but from the perspective of the other players in the game watching what their opponent
does it may be that there are many points at which Player 1’s move looks random.
If you wanted to describe a general mixed strategies for Player 1 in this game you would
write down a probability distribution (p, q, r, 1 − p − q − r) where p is the probability (R, R)
is played, q is the probability (R, F ) is played, r is the probability (F, R) is played, and
1 − p − q − r is the probability (F, F ) is played.
The second way of including randomness is to allow for the instructions themselves to
have random elements. That is at each place where the player must move the instruction
book tells the player how to randomize at that point. For example such an instruction
book for Player 1 might be: At first node with probability 12 play Raise. At second node
with probability 14 play Raise. These books with random instructions are called Behaviour
Strategies.
Definition 3. A Behaviour Strategy for a player is an instruction book that has random
instructions. A.#Perfect#Information#Games#in#Extensive#Form6
Here’s our example of this random instruction book in action.
1’s#Behaviour#strategy6 1
6
Fold'½'' Raise'½'
2 2
(21,1)' (1,21)'
(0,0)'
1
Fold'¾''' Raise'¼''
(21,1)' (2,22)'
If you wanted to write down a general behaviour strategy for Player 1 in this game. You
would write the pairs (p, 1 − p) and (q, 1 − q) where (p, 1 − p) is the randomization at the
first decision Player 1 must take and (q, 1 − q) is the randomization at the second decision
she must take.
2: GAMES IN EXTENSIVE FORM 7
Fold'' Raise'
2 2
(11,1)' (1,11)'
(0,0)'
(0,0)'
B.#Incomplete#Information#Games#in#Extensive#
Form7
To get over this we make the following addition to the picture.
2#does#not#see#1’s#action71
Informa7on'Set'
(players#move#simultaneously)7
Fold'' Raise'
2 2
(11,1)' (1,11)'
(0,0)'
(0,0)'
8 MARTIN CRIPPS
This indicates that 2 does not know what 1 has done when she chooses between Raise and
Fold.
Often we will use dashed lines to indicate information sets—this is done in all the
following examples.
Information sets can be used to describe situations where players move simultaneously.
But there are many more subtle things that they can be used to represent
(1) Information Sets can be used to describe situations where a player’s past actions
are hidden from others (this is like the simultaneous move tree).
(2) Information Sets B.#Incomplete#Information#Games#in#Extensive#
can be used to describe situations where a player has private
information. For example, Player 2 knows the outcome of a coin toss but Player 1
Form7
does not. Hidden#information:#player#2#knows#the#outcome#of#a#coin#
toss#but#player#1#does#not7
0
1/2%% 1/2%
2 2
(21,1)%
(21,1)% (2,22)% (21,1)%
(3) Information Sets can be used to describe situations where a player forgets what
B.#Incomplete#Information#Games#in#Extensive#
they knew previously. For example, Player 2 knows the outcome of a coin toss and
Form7
then forgets it. Hidden#information:#player#2#forgets#they#the#outcome#of#a#
coin#toss7
0
1/2%% 1/2%
2 2
(21,1)%
(21,1)% (2,22)% (21,1)%
When we introduce information sets we need to change the definition of what a pure
strategy is:
2: GAMES IN EXTENSIVE FORM 9
Definition 5. A pure strategy for an incomplete information extensive form game is just
an instruction for every information set where it must make a move. (NOT instructions
for every node.)
If a player does not know which node at an information set they actually are at they are
forced to choose the same action at every node in that set. For example, in the simultaneous
move game above Player 2 only has two pure strategies Fold or Raise. But in the game
without the information set Player 2 has four pure strategies (Fold,Fold), (Fold,Raise),
(Raise,Fold), (Raise,Raise), that is, they can act in a way that depends on what Player 1
did.
3. Nash Equilibrium
We have already defined what a Nash equilibrium is: At a Nash equilibrium each player’s
strategy is a best response to the other players’ strategies. This seems to be a good notion
of equilibrium in strategic form games, but (as we will see) it is not such a great notion
of equilibrium in extensive form games. Remember the test for a Nash equilibrium is:
fix everyone else’s strategy then check whether my strategy is optimal for me given that
everyone else is committed to playing their strategy.
To understand why a Nash equilibrium might not be a good way of describing equilibrium
in these games let us look at two Nash C.#Nash#Equilibrium0
equilibria of the same game. Here is a simple two-
stage game and one Nash equilibrium of it.
A#Nash#Equilibrium.0 1
Fold'' Raise'
2 2
If we fix Player 2’s strategy (in green) you can see that changing 1’s strategy will reduce her
payoff to −2 from −1, so certainly 1 is doing the best she possibly can given 2’s actions.
Now let us test 2’s strategy. If we change what 2 does at the right decision it has no
effect on her payoff (she gets 1), so any action here passes the optimality test of a Nash
equilibrium. If we change 2’s action at her left decision her payoff goes down to zero from
1, so again she is acting optimally here. I hope this convinces you that this is a Nash
equilibrium.
Now we will argue that the above is not a particularly sensible equilibrium. Consider
again why Fold is optimal for Player 1. This is because she believes she will get −2 if she
plays Raise. Is this a sensible belief? The answer to this is yes under the assumption that
Player 2 plays the green actions, but is it reasonable to expect that Player 2 will play the
green action? The answer to this is no. Player 2 is threatening to play Raise if Player 1
10 MARTIN CRIPPS
Raises. However, this threat is not something Player 2 would actually want to carry out—it
is a non-credible threat. Because if 1 actually played Raise, the best thing for Player 2 to
do would be to Fold (they would get −1 rather than −2). Thus the problem with this NE
is that Player 2 is threatening to do something that they are not actually going to do and
as a result of the NE assumption Player 1 believes this threat.
We have just argued that this is a bad kind of equilibrium, so is there another NE of
this game that does not have this problem? Here is another NE of the same game that
does not rely on a non-credible threat.
C.#Nash#Equilibrium0
Credible#Threat.0 1
0
Fold'' Raise' This'then'changes'1’s'
0 behaviour.'
0
2 2
0
Fold'' Raise' Fold'' Raise'
0
0
(0,0)' (11,1)' (1,11)' (12,12)'
0
Once#player#1#realizes#this#threat#is#non=credible#she#is#
be>er#off#raising#too0
This is another Nash equilibrium and at this Nash equilibrium no non-credible threats
are made. Can we always find such a Nash equilibrium? The answer is yes and what
is more there is a very nice process to find these good Nash equilibria and it is called
backwards induction. This always works in games of perfect information (i.e. games
without information sets), so in these games we can always find Nash equilibria that only
rely on credible threats.
Example2 1
2
Fold'' Raise'
2
2
2 2
2
Fold'' Raise' Fold'' Raise'
2
2
(11,1)' (1,11)'
2 (0,0)'
1
Start#at#the#last#move#and#figure2 Raise'
Fold''
out#what#is#optimal#there2
2
(11,1)' (2,12)'
(As −1 < 2 Player 1 prefers Raise at the last node.) Then you determine the optimal
decisions at all the other final nodes.
D.#Backwards#Induction2
Example2 1
2
Fold'' Raise'
2
2
2 2
2
Fold'' Raise' Fold'' Raise'
2
2
(1,11)'
2 (0,0)' (11,1)'
1
Do#all#last#moves2 Raise'
Fold''
2
(11,1)' (2,12)'
Example2 1
2
Fold'' Raise'
2
2
2 2
2
Fold'' Raise' Fold'' Raise'
2
2
(0,0)' (11,1)' (1,11)'
2 1
Now#do#the#second#last#move.2 Raise'
Fold''
2
(11,1)' (2,12)'
Example:2 1
2
Fold'' Raise'
2
2
2 2
2
Fold'' Raise' Fold'' Raise'
2
2
(0,0)' (11,1)' (1,11)'
2 1
Finally#do#the#first#move2 Raise'
Fold''
2
(11,1)' (2,12)'
• Backwards induction will not work in games with information sets, so it can only
work in perfect information games.
• Reinhart Selten generalized the idea of backwards induction to create the idea of
subgame perfection and got the Nobel prize for it (we will study this in the next
section).
4.1.1. Investment by multinational firms. In this model we think of a firm first deciding
whether to invest overseas or not. Then the government of the destination country de-
D.#Backwards#Induction#2
cides whether to expropriate the investment or not. The country would like to commit
to not expropriating but, in the simple version of this game below, this is not a credible
undertaking. Given#expropriation#it#is#optimal#not#to#invest.2
(50,50)
DC
MNF
Expropriate
(-20,80)
Do not invest
(0,0)
A#(Bad)#Nash#Equilibrium31
Fold'' Raise'
(1,1)' 2
L'' R'
1 1
In this game Player 1 folds because they believe they will get a payoff of 0 if they raise.
Player 2’s action genuinely has no effect on their payoffs so R is optimal for them. Hence
the situation above passes the test for a Nash equilibrium. However, if Player 2 got to
actually make a move in this game, then going R gets them 1 while going left gets them
2. It looks like Player 2 is making a threat they wouldn’t actually want to carry out. We
would like to rule this situation out by doing something like backwards induction. But
that’s not going to work in this game as we cannot work out what Player 1 would do at
the very last move—it might be optimal for 1 to do action a if she thought she was at the
left node but it might be optimal for her to do action b if she thought she was at the right
node.
Reinhard Selten suggested that we treat the events after Player 1 moves Raise as a
separate independent game, or a subgame, of the original game. He suggested that what
players do on this separate independent game must be a Nash equilibrium of that game.
In this example, this is a game where Player 2 and Player 1 simultaneously move. The
2: GAMES IN EXTENSIVE FORM 15
L R
a (2,2) (0,1)
b (-1,1) ( 12 ,0)
What the players are doing at the bad NE of this game is (a, R). If you were given just
this game to analyse you would predict (a, L) because it is the unique Nash equilibrium
and can be found by iteratively eliminating strictly dominated actions. Selten said that
we should expect the players to play a NE of this game, so if we replace the (a, R) actions
with the Nash equilibrium of the subgame we get the following picture.
E.#Subgame#Perfection3
3 1
3
Fold'' Raise'
3
3
(1,1)' 2
Once3
L'' R'
we#have#solved#this3
last#game#it#is#easy3 1 1
what#player#1#will3
do.3 (2,2)' (11,1)' (0,1)' (1/2,0)'
Here there is also a dashed line around the subgame we solved separately (and some text
I have been unable to crop). Once we have solved this subgame we can easily figure out
what Player 1 will do at the initial move and find a subgame perfect equilibrium of the
entire game. If Player 1 plays Fold they will get 1 but if they play Raise and the NE of
the subgame is then played, then Player 1 will get 2. Thus this SPE has player 1 Raising.
Summary: The way of finding a subgame perfect equilibrium is to find a Nash equilibrium
on the smallest subgames and then work backwards through the treat finding a Nash
equilibrium on subgames as you go.
There is a problem with subgame perfect equilibria in that they may not be unique.
This occurs because the subgames may have many Nash equilibria. To see this consider a
slightly different version of our game.
16
E.#Subgame#Perfection3
MARTIN CRIPPS
3 1
3
Fold'' Raise'
3
3
(2,2)' 2
L'' R'
1 1
The subgame of this extensive form can be represented as the strategic form below.
L R
a (3,1) (0,0)
b (0,0) (1,3)
This strategic form has 2 pure Nash equilibria at (3, 1) and (1, 3). So which one do we
choose to put in the game to find the subgame perfect equilibrium (SPE)? Well we can
choose either of these.
If we choose the (3, 1) NE E.#Subgame#Perfection3
on the subgame we get the SPE below.
3 1
3
Fold'' Raise'
3
3
(2,2)' 2
L'' R'
1 1
3 1
3
Fold'' Raise'
3
3
(2,2)' 2
L'' R'
1 1
It is clear that one can choose the NE to get very different behaviour from Player 1 at
the first move. But these are both perfectly fine SPE’s.
3: MONOPOLY
MARTIN CRIPPS
In this section we resit some models you probably studied last year so we can warm up
for the study of more complex models of firm behaviour.
To maximize profit (and choose the optimal price-output combination) we differentiate and
set the slope of profit equal to zero.
P = qP (q) − C(q, w)
dP dp dC
=q + P (q) − =0
dq dq dq
Marginal Revenue − Marginal Costs =0
Notes: the formula for marginal revenue does not equal the price. This is because to sell
more a monopolist has to cut the price on all units it sells not just the additional unit.
Thus
dP
MR = q + P (q) < P (q)
dq
1 MC
Optimal Price
AC
Demand
MR
Optimal Quantity
15
(The arrow indicates the profit per unit.) As the demand moves around it is not clear how
the price will change. It is possible that an increase in demand reduces the price.
An#increase#in#demand#can#reduce#price.#/
/
/
/ New
New
/ Demand
MR
/
/
/ MC
/
/
AC
/
/
/
/
/ Old MR Old Demand
Note:#Here#the#new#demand#is#much#fla8er#and#quantity#has#gone#up./
/ 17
/
You see this in DVD’s where low demand DVD’s often are quite expensive but very popular
family films are much cheaper.
It is also possible that an increase in demand reduces the price.
3: MONOPOLY 3
An#increase#in#demand#can#increase#price.#.
.
New
. Demand
. New
MR
.
.
.
MC
.
. AC
. Old Demand
.
. Old MR
Note:#Here#the#new#demand#is#steeper..
. 19
You see this in fashion when some new fad becomes popular and simultaneously expensive.
To understand what’s going on here we need to take the condition for profit maximisation
and reinterpret it in terms of pricing:
dp dC
q + P (q) =
dq dq
q dp dC
P (q) +1 =
p dq dq
" #
p dq −1
dC
P (q) ( ) +1 ==
q dp dq
Hence we get
dC
dq MC
P (q) = −1 = 1
p dq 1+
q ( dp ) +1 Price Elasticity
It is this rule that tells us about pricing:
• As marginal cost goes up so do prices.
• If demand becomes more elastic (the elasticity moves from −3 to −4 for example)
so prices go down.
This results in the following typical properties: Steeper demand curves have higher prices
and flatter demand curves have lower prices. This is what is going on the in the previous
examples.
1.1. Monopoly Power. Pure monopoly is rare but firms do face downward sloping de-
mand curves. As one firm varies its price there will be a change in the demand the firm
experiences.
The individual firm’s demand is more responsive to price than the market demand, but
as long as demand does not entirely vanish when it raises its price the firm enjoys some
monopoly power.
4 MARTIN CRIPPS
One way of measuring monopoly power is to consider the share of the price that goes in
profit.
Price − MC
Lerner’s Measure of Monopoly =
Price
If the firm is maximising profits and we substitute in for the profit maximising price from
above into this measure we get that Lerner’s measure is equal to = −1/Elasticity of demand.
So the question of what determines a firm’s monopoly power amounts to the question:
What determines the firm’s elasticity of demand?
1.2. Two Part Tariffs. This is a topic we will revisit later in this course but also should
appear in the section on Monopoly.
A two-part tariff is a common form of pricing where the price paid by the consumer for
buying Q units is P = A + pQ, where A is a fixed fee (connection charge, membership fee,
admission charge) and p is the price for units. Clearly, the average price = (A + pQ)/Q
declines in Q. Effectively this is a quantity discount, so high-usage customers pay less per
unit than low-usage consumers. On the other hand it seems to be a deterrent to consuming
small numbers of units. You see two-part pricing in: Theme Parks, Utilities, all you can
eat restaurants, razors. . .
We have talked about how one might choose a price optimally when there is no fixed
fee, but how does this decision change in the presence of the fee. To understand the best
way of setting the fee A it is easiest to think of the case where there is only one consumer
buying units of your good. In this case the area below the demand curve of the individual
and above the price represents the value the individual is getting from the good but is not
paying for. Here is the example of a demand function and a per unit price.
3: MONOPOLY 5
Example:Telephone-service3
3
3
120
price
(cents per minute) Q=120-P Surplus = $50
charge
per minute 20
0 100 120
quantity (minutes/month)
33
The area above the price but below the demand is a triangle base 100 height 100 so it has
an area 5000 or $50. This represents the value the consumer is getting at this per unit
price that they are not paying for. This is the maximum fixed fee the seller could charge.
In all they would make 20 × 100 on per-unit sales and then 5000 in a fixed fee. The revenue
the firm gets is the entire area under the demand curve between q = 0 and q = 100. Thus
we can now address the optimal choice of a two part tariff by appreciating that in this case
the firm’s revenue is the area under the demand not just price times quantity.
Z q
Total Revenue = P (v)dv
0
Z q
d
Marginal Revenue = P (v)dv = P (q)
dq 0
So to maximise profits you set M R = M C, which implies P (q) = M C or that price is set
to equal marginal cost—just like a perfectly competitive firm would.
1.3. Price Discrimination. In general setting only one price for a product is suboptimal
because there are usually customers who are willing to pay more. The practice of setting
several prices for the same product is called price discrimination.
Price Discrimination occurs when a firm sells the same good at different prices to different
customers. Firms do this to try to capture part of the consumer surplus; this is not a bad
thing as generally it increases supply. Here are some examples of it
• Airlines(Business/Vacation)
• Hotels(Business/Vacation)
• Drugs (Generic/Branded)
• Telephone Companies (Fixed Fee and Call charges)
• Service provides Fixed fee and per unit cost (IBM/Xerox)
• Manufacturers(Offer high initial prices and lower ones later)
• Businesses(offer student discounts)
• Manufactures (Retail/Wholesale)
6 MARTIN CRIPPS
1.3.1. First-Degree Price Discrimination. Here a customised price is set for each individual
customer and it may be observed in car dealers, dotcoms, telephone companies, college
tuition. In practice it is very unlikely, because firms cannot find out customers true value.
This theoretical extreme would and it maximise profit if it were feasible.
(If a firm is able to do this, then the revenue is the area under the demand curve. The
revenue obtained from selling one more unit = price. Thus, it is optimal for the firm to
produce the quantity at which M C = p.)
1.3.2. Second-Degree Price Discrimination. This describes a situation where a firm is able
to charge a price that depends on the quantity purchased (e.g. quantity discounts). Ex-
amples: Block Pricing by Power Generators, minutes on your phone, two-part tariffs.
1.3.3. Third-Degree Price Discrimination. This describes everything else. For example
where the firm divides a market into 2 broad segments and charges these segments different
prices. If a firm supplies two markets A and B with two independent plants, then to
maximise profits it wants to set the prices in the segments in the way that will maximize
its profit. That is
MCA MCB
PA = 1 , PB = 1
1+ 1+
Price ElasticityA Price ElasticityB
To see where this result comes from, suppose that the firm supplies output from one plant to
two separate markets, then there is a trade-off between supplying one market and another
because it increases costs.
Profits = Π = qA p(qA ) + p(qB )qB ?C(w, qA + qB )
Maximizing:
dΠ dΠ
0= , 0=
dqA dqB
Or
dC dC
= p(qA ) + qA p0 (qA ) = p(qB ) + qB p0 (qB )
dq dq
3: MONOPOLY 7
Rearranging
dC dC
= pA [1 + (qA /pA )p0 (qA )] = pB [1 + (qB /pB )p0 (qB )]
dq dq
and so. . .
MC MC
= pA = pB .
C.#State#Control#of#Monopoly!"
1 + (ElasticityA)
−1 1 + (ElasticityB )−1
1.4. Regulation ofPrice#Regulation!of!a!monopoly:!"
a Monopoly. The most common form of regulation a firm faces is a
maximum price, or a•If!the!price!is!regulated!to!be!no!more!than!P(max)"
price cap. We will see that this will tend to increase the output of a
"
monopoly and decrease its price. Here is a picture of a monopoly and a potential regulated
price (P (max)). Note that this regulated price is less than the price it would charge if it
set M R = M C.
Demand
MC
P(max) AC
MR
State%Control%of%Monopoly!"
53
The regulated price means that the firm’s M R curve changes. If the firm is selling less
than the maximum it can sell at the regulated price, then it can always sell one more unit
Price%regulation%of%a%monopoly:%6
•Beyond!this!point!MR!is!as!it!was!before."
and get the regulated price for it. Thus the firms M R is flat and is equal to the regulated
•What!is!the!optimal!quantity!now?"
price, until the regulated price hits the demand curve. At this point the M R is exactly
"
what it was before. This is shown in the figure below.
Demand
MC
P(max) AC
MR
56
It should be clear from this figure that M C = M R at the vertical section of the red M R
curve. Thus the amount of output the firm has produced has increased and prices have
fallen as a result of this regulation.
4: STATIC OLIGOPOLY
MARTIN CRIPPS
In this topic we take the tools of strategic form games (in particular Nash equilibrium
and dominance) to analyse models of industries where there are many firms who are in
competition with each other. There is no single “correct” model of these kinds of industry.
Here we will look at four models of Oligopoly, but there are others you should be aware of.
1. Cournot
To understand how Cournot’s model of oligopoly works it is best to work through an
example. Before doing this we imagine a world where firms are producing exactly the
same product (homogenous product) and the decision they must take is how much output
to produce. By producing more output they drive down the market price for themselves
and all other firms in the industry. (We say this negative effect of one firm on others in
the industry is a “negative externality”.) One interpretation of this model is it describes
competition in the long run. That is when firms are deciding what factories to build. It is
then, presumably, that they make their quantity decisions.
1.1. Example of Two-Firm Cournot Oligopoly. In the example we imagine there are
only two firms in the industry each producing exactly the same good (water or oil for
example). The demand curve for their product is P = 30 − Q, where Q is the total output
produced by all the firms and P is the market price. Thus we could write P = 30−(Q1 +Q2 ),
where Q1 and Q2 are the individual output of firm 1 and firm 2 respectively. Each of the two
firms has the same costs of production with a constant Marginal Cost of 6: C(Q1 ) = 6Q1
and C(Q2 ) = 6Q2 . (A quicker way of writing this is C(Qi ) = 6Qi for i = 1, 2.)
With this information it is possible to write down the firms’ profits (or payoffs) in this
game:
π1 (Q1 , Q2 ) = P Q1 − C(Q1 ) = (30 − (Q1 + Q2 ))Q1 − 6Q1 ,
π2 (Q1 , Q2 ) = P Q2 − C(Q2 ) = (30 − (Q1 + Q2 ))Q2 − 6Q2 .
What is different from Monopoly is that each firm’s profit depends on the output of another
firm.
To summarise the game: Each firm chooses and output level Qi ≥ 0. These choices are
made simultaneously. The payoffs these choices are given by the firms’ profits π1 and π2
in the above equation.
Our tool to analyse this situation will be to find a Nash equilibrium. And because of
the large number of actions the firms have the method of finding the Nash equilibrium will
1
2 MARTIN CRIPPS
be to find the firm’s best responses. We called this “Method 4 The Intersection of Best
Responses”.
1.1.1. Finding the Best Responses. Here are two ways of calculating Firm 1’s best response
to a given output of Firm 2. First write down firm 1’s profit.
π1 (Q1 , Q2 ) = (30 − (Q1 + Q2 ))Q1 − 6Q1
Observe that this is a quadratic function of Q1 os its maximum can be found by differen-
tiating and setting equal to zero.
∂π1
= 30 − (2Q1 + Q2 ) − 6 = 24 − Q2 − 2Q1 = 0.
∂Q1
Solving this for Q1 gives
1
24 − Q2 = 2Q1 , orQ1 = 12 − Q2 .
2
Thus firm 1’s optimal choice of output is a linear and decreasing function of firm 2’s output
choice.
A second and more Economist-y approach to this would be to say that Firm 1’s Revenue
= P Q1 = [30 − Q1 − Q2 ]Q1 . So we differentiate this to get its Marginal Revenue
dR
MR = = 30 − 2Q1 − Q2 .
dQ1
We also know that the firm’s Marginal Cost = 6 from above. So to find where the firm
maximises profit we would set MC equal to MR, that is,
1
MC = 6 = 30 − 2Q1 − Q2 = MR or Q1 = 12 − Q2 .
2
This tells us firm 1’s best output as a function of firm 2’s output. This relationship is
important and is often plotted by economists. Here’s a picture
In#pictures:-
Q2"
Q1"="12"–"0.5"Q2"
Q1"
(2) The intercepts are informative. When Q2 = 0 and firm 1 produces its optimal
response it is alone in the market. Thus Firm 1’s optimal response to Q2 = 0 is
the monopoly level of output. In this example 12 is the output of a monopolist.
(3) The intercepts are informative (again). When Q2 is very high, Firm 1 wants to
produce zero output, because it makes zero profit. (In fact when Q = 24 this is
true.) Zero profit is actually a characteristic of Perfect Competition. Thus, in this
example Q = 24 is the perfectly competitive output.
Notice that everything is about Firm 1 is true for Firm 2. So when we try to calculate
Firm 2’s Best Response we will get a very similar kind of equation:
1
Q2 = 12 − Q1 .
2
Just the names of the outputs have been swapped.
1.1.2. Finding the Cournot/Nash Equilibrium. We now have two reaction functions of the
firms: Q1 = 12 − 12 Q2 and Q2 = 12 − 12 Q1 . If we are going to find a Nash equilibrium of
this game we are going to!Best
findresponse
where they intersect.
functions This
intersect is usually
at the Nash called a Cournot, or
Cournot-Nash equilibrium, because
equilibrium"Cournot found this solution to the model many many
years before Nash was even born. Here’s the picture of the two reaction functions:
Q2#
Cournot#(Nash)#Equilibrium#
Q2#=#12'0.5Q1#
Q1#
15#
To find the Cournot equilibrium we must substitute one reaction function into the other
1
Q1 = 12 − Q2
2
1 1
Q1 = 12 − [12 − Q1 ]
2 2
3
Q1 = 6
4
Q∗1 = 8
Performing a similar calculation for Firm 2 we get Q∗1 = Q∗2 = 8.
There is a short cut that we can use in this particular case because the firms have the
same costs. In this case we know the equilibrium will be symmetric Q∗1 = Q∗2 . So instead of
substituting firm 2’s reaction function into firm 1’s reaction function we can just substitute
4 MARTIN CRIPPS
This writes the relationship between prices and output in three identical way. For our
purposes the last is the most useful. It says price equals D minus
P firm i’s output and the
output of all the other firms. It will be useful to write Qi = j6=i qj as the output of all
firms apart from firm i.
Now we are in a position to write down firm i’s profit.
πi (qi , Qi ) = pqi − cqi = (D − Qi − qi )qi − cqi .
To find Firm i’s best response we want to maximise this by choosing the best qi . So, as
above, we differentiate and set equal to zero.
∂πi
= D − Qi − 2qi − c = 0.
∂qi
Setting this equal to zero gives qi∗ = (N − Qi − c)/2. If we do this for each firm we will get
N equations in N unknowns:
q1∗ = (D − Q1 − c)/2, q2∗ = (D − Q2 − c)/2, ... ∗
qN = (D − QN − c)/2
This can be solved directly, but it is much easier to use the trick we talked about in the
earlier section. Here all the firms are the same and have the same costs. We would expect
that the Nash equilibrium had all the firms producing the same output q1∗ = q2∗ = · · · = qN
∗ .
4: STATIC OLIGOPOLY 5
So let us substitute this into Firm 1’s reaction function. With this assumption, the output
of all firms apart from firm 1 can Q1 = N ∗ ∗
P
i=2 i = (N − 1)q1 can be calculated. Hence we
q
get
q1∗ = (D − Q1 − c)/2 = (D − (N − 1)q1∗ − c)/2.
This solves for q1∗ to give
D−c
qi∗ = .
N +1
Hence as N increases each firm’s output tends to 0, there are more and more firms and
each becomes vanishingly small relative to the size of the market. We would hope this
would look something like perfect competition, and we can check this by finding out how
the price behaves as N increases. The price will satisfy
X D−c Nc + D
p∗ = D − qi∗ = D − N = →c as N → ∞.
N +1 N +1
Here the price tends to Marginal Cost as the number of firms grows. Thus Cournot
competition starts to look a lot like perfect competition when there are a lot of firms.
2. Bertrand Competition
This model of competition is similar to Cournot’s in that it assumes that firms are
producing exactly the same product. However, it assumes that firms choose the price of
their product (not how much to produce). This model of price competition is a good
description of what goes on in financial markets or of the price wars that many industries
experience from time to time. One interpretation of this is that it describes what happens
in the short run, rather than the long run.
To be precise there are three assumptions in Bertrand competition:
(1) The lower-priced firm always claims the entire market. (Firms produce identical
products.)
(2) All competition is in prices.
(3) If the firms set equal prices they will share the market.
Again it is easiest to understand Bertrand competition by thinking about an example.
2.1. An Example of Bertrand Competition. There are two ferry companies (it is
more usual to consider airlines but I am quite find of boats). They serve the same route
offering identical service, the only difference between the two companies is the price they
charge. The cost per customer is 30 for both companies. There are 1000 customers who are
willing to pay up to 50 to make the crossing. Company 1 charges a price P1 and company
2 charges P2 .
First let us write down the profits of Firm 1. If it is the low price firm it gets all 1000
customers and has profit
π1 = 1000(P1 − 30), P1 < P2 .
If it has the same price as Firm 2 it gets half of 1000 customers and has profit
π1 = 500(P1 − 30), P1 = P2 .
6 MARTIN CRIPPS
If it has a higher price than Firm 2 it gets no customers and has profit
π1 = 0(P1 − 30) = 0, P1 > P2 .
B.#Nash#Equilibrium:#Bertrand#Competition8
Here is a picture of this function (remember M C = 30).
Firm#1;s#Profit8
MC P2
The most important thing we notice from this is a problem. Firm 1’s profit is highest if it
just undercuts firm 2’s price. But it also wants to charge as high a price as possible. Thus
what Firm 1 would like to do is to charge the highest30# price just below firm 1’s price, but no
such price exists. Thus, in this game (with continuous prices) the best response function
of firm 1 does not exist!
The question then is: can we find a Nash equilibrium if we cannot find where the best
responses intersect? Well the answer is yes, by telling a story. Suppose, the firms set equal
prices of 50 and share the market (which is clearly a good situation). The profit of each
firm is 500(50 − 30) = 10, 000. If firm 1 cut its price to 49, then it can attract all the
customers and make profit 19, 000 = 1000(49 − 30). In response firm 2 can undercut firm 1
by asking the price 48 and attracting the whole market. This process continues until both
firms are charging the price 30 and making zero profit. At this point no firm benefits by
undercutting their rival and the firms do actually have a best response. Summary: The
Nash equilibrium is at (P1 , P2 ) = (30, 30), although there is no best response function.
2.2. General Properties of Bertrand Competition. Here is a list of points that should
be noted:
• Pure price competition drives oligopoly to look very much like perfect competition.
The Nash equilibrium of the game has the firms setting prices equal to marginal
costs. This is why we believe markets with price competition such as financial
markets may be quite efficient.
• Price competition does not expand the market in the above example—the demand
was always 1000 for low prices. As firms cut prices, one would usually expect more
customers to want to buy the good. This expansion of the size of the market will
increase the temptation of the firms to undercut their rival’s price.
• When prices are not allowed to vary continuously the problem of the non-existence
of a best response goes away (see the next section).
2.3.1. Integer Prices. A lot of the problems with Bertrand competition came from the fact
that firms were allowed to continuously vary their prices. If we assume firms are only able
to charge whole number prices then there is no problem with finding the biggest price less
than some number. It also increases the number of Nash equilibria of the game. Here are
the firms profits for the prices 30, 31, 32, 33, in the model of ferry competition above.
P2 = 30 P2 = 31 P2 = 32 P2 = 33
P1 = 30 (0,0) (0,0) (0,0) (0,0) ...
P1 = 31 (0,0) (500,500) (1000,0) (1000,0) ...
P1 = 32 (0,0) (0,1000) (1000,1000) (2000,0) ...
P1 = 33 (0,0) (0,1000) (0,2000) (1500,1500) ...
.. .. .. .. ..
. . . . .
If you apply the underlining method to this game you will find that it has 3 Nash equilibria:
(P1 , P2 ) = (30, 30), (P1 , P2 ) = (31, 31), (P1 , P2 ) = (32, 32). This increase in the number of
Nash equilibria arises because now you really have to make a big change in your price if
you are going to undercut your rival. This big change in the price might hurt your profits
more than the increase in customers you experience from being the low-price firm.
2.3.2. Price Guarantees. If firms advertise deals like “if you find this good cheaper any-
where else we will refund twice the difference” then there is a strange effect. Suppose the
firms set prices (40, 41) where would the customers go? If they go to the firm with the
sticker price of 40, that is what they will pay. If they go to the firm with the sticker price of
41 they can claim a refund of twice the difference in prices and in fact pay only 39. Thus all
the customers would prefer to go to the higher price firm and claim a refund. The sensible
response of the firm setting the low price (40) is to raise its sticker price, so the customers
come to it and claim the refund. If such guarantees are in place we would expect prices to
increase rather than decrease. These guarantees are collusive, although they look like they
are good for the consumer.
2.3.3. Capacity Constraints. If firms are unable to fit all 1000 customers on their ferry it
never makes sense to cut prices down to Marginal Cost. Instead prices seem to go around
in circles. These circles are called Edgeworth cycles.
Consider the following slight change in our example. Suppose that only 800 customers
will fit on a ferry. So that when you are the low-price firm your profits are 800(P − 30)
and when you are the high-price firm your profits are 200(P − 30). The high price firm
always gets to serve those who cannot get on the low-price ferry. Suppose now that firm
1 sets the price of 35. Firm 2 can undercut this and set the price of (say) 34 making the
profit 800(34 − 30) = 3200. Or firm 2 can embrace being the high-price firm and charge
the price of 50 and make the profit 200(50 − 30) = 4000. Clearly the best response is to
set the price P2 = 50. Now (of course) firm 1 will respond by raising prices to P1 = 49 and
competition will drive prices back down again to 35 where the cycle begins again.
8 MARTIN CRIPPS
3.1. Example of Price Competition with Differentiated Products. There are two
firms producing different goods. P1 is the price chosen by Firm 1 and P2 is the price chosen
by Firm 2. We will first describe the demand functions of each firm.
Firm 1 : Q1 = 12 − 2P1 + P2
Firm 2 : Q2 = 12 + P1 − 2P2
Notice that each firm’s demand is decreasing in its own price but increasing in its rival’s
price—they are producing substitute goods and are in competition with each other. We
will ignore costs here and suppose that all the costs are fixed costs. Fixed Costs = 20.
The first step in defining this game is to write down the Firms’ actions. These are the
prices they choose (P1 , P2 ). The next step is to write down the Firms’ payoffs or profits.
Firm 1’s Profit : = P1 Q1 − Cost = P1 (12 − 2P1 + P2 ) − 20
Firm 2’s Profit : = P2 Q2 − Cost = P2 (12 − 2P2 + P1 ) − 20
This completes the formal description of the game that is being played here.
Now we must find a Nash equilibrium of this game and again we will use the Reaction
Function method. We begin by finding the best P1 for firm 1 given it knows the price of
firm 2. To do this we maximize firm 1’s profit by differentiating and setting equal to zero.
dπ1
= 12?4P1 + P 2 = 0
dP1
This solves to give
1
P1 = 3 + P2 .
4
A similar process for firm 2 will also find its optimal choice of P2 as a function of P1 :
1
P2 = 3 + P1 .
4
Note here these reaction functions are upward sloping. As my rivals price goes up it is
optimal for me to increase my price too. (I can still undercut by raising my price and
an increase in my rivals price has a positive effect on my demand.) To find the Nash
equilibrium (P1∗ , P2∗ ) we must substitute one reaction function into the other.
1 1
P1 = 3 + (3 + P1 ).
4 4
C.#Nash#Equilibrium:#Differentiated#Product#
4: STATIC OLIGOPOLY 9
Duopoly<
This solves to give (P1∗ , P2∗ ) = (4, 4). Here is a picture of what we have just done.
The#Firms’#Reaction#Curves#or#Optimal#Responses<
#<
#< P2#
P1 = 3 + 0.25P2
#<
#<
#<
#< P2 = 3 + 0.25P1
#<
#<
#<
#
<
< P1#
<
The#Nash#equilibrium#is#at#(P1,P2)=#(4,4).##<
47#
4. A Model of a Single-Unit Auction
All the previous models were games where a few sellers compete to sell to a set of buyers.
Now we will look at the reverse position where a few buyers compete to acquire a good
from a single seller.
First we will describe the buyers. The buyers? values for the good are written as
(v1 , v2 , . . . , vN ) where buyer I’s value, vi is in the interval 0 ≤ vi ≤ 1. We will suppose
these values are random and that the buyers only know their value but not the value of
the others. The seller does not know the values (v1 , v2 , . . . , vN ) and so has a random or
unknown demand curve.
How the (v1 , v2 , . . . , vN ) are determined has a big effect on the nature of competition
that among the buyers. Here are some different assumptions that might be made:
Independent Symmetric Private Values: vi is drawn independently from the den-
sity f (v) on [0, 1]. All buyers have the same distribution of values for the good.
Independent Private Values: vi is drawn independently from the density fi (v) on
[0, 1]. Buyers have distinct views about the good.
Here is the order of events in the auction
(1) The players observe their own values and no-one else?s.
(2) Then they submit a bid.
(3) The rules of the auction determines payoffs.
Here a strategy for a buyer is to describe how they should bid for each different value they
observe. Thus players’ strategies are bidding functions that takes the player’s value and
maps it to a bid.
#5 #b#
bi#(.)#
bi#(vi)#
v##
0# vi## 1#
We have to find a whole function to describe a player’s equilibrium strategy. To make this
easier we will only look for certain kinds of equilibria. That is, equilibria where bids are
strictly increasing functions of values.
4.1. Equilibrium in Second-Price Auctions. The rules of the auction determine who
wins and who pays what. The easiest and simplest set of auction rules to analyse are
second-price auctions. These are auctions where the person submitting the highest bid
gets the object, but they pay a price equal to the second highest bid (not their own highest
bid). In an auction where the price paid is the second highest bid, the strategy bi (vi ) = vi
(that is submit a bid equal to your value), weakly dominates all other strategies. This is a
famous result due to Vickrey and is much used in Economics.
We now explain why this is true.
We first consider the possibility of bidding above your value. That is, overstating how
you feel about the good. Suppose you have a value vi and consider your payoffs for a bid
b0 > vi . Your payoff only depends on the highest bid from the other players call this B. We
will deal separately with the cases where: (1) the highest bid from everyone else is above
your new contemplated bid B > b0 . (In which case you always lose the auction whether you
bid truthfully or exaggerate to b0 ). (2) The highest bid from everyone else falls between
your value and your increased bid b0 > B > vi . (In which case bidding truthfully causes
you to lose the auction and get zero while bidding b0 causes you to win the auction and
pay B > vi so you get a negative payoff.) (3) The highest bid from everyone else is below
your value vi > B. (In which case you win the object whether you bid vi or b0 and the
price you pay B is independent of your bid. This gives the matrix of payoffs below. You
can see that the top row is weakly dominated by the bottom row.
B > b0 B ∈ (vi , b0 ) B < vi
0
Bid b Lose = 0 Win = vi − B < 0 Win = vi − B > 0
Bid vi Lose = 0 Lose = 0 Win = vi − B > 0
Now we consider the possibility of bidding below your value. That is, understating how
you feel about the good. Suppose you have a value vi and consider your payoffs for a bid
4: STATIC OLIGOPOLY 11
b00 < vi . We will deal separately with the cases where: (1) the highest bid from everyone
else is above your value B > vi . (In which case you always lose the auction whether you
bid truthfully or understate your bid). (2) The highest bid from everyone else falls between
your value and your understated bid b00 < B < vi . (In which case bidding truthfully causes
you to win the auction and pay a price B < vi giving positive profit while bidding b00 causes
you to lose the auction and get nothing.) (3) The highest bid from everyone else is below
your understated bid b00 > B. (In which case you win the object whether you bid vi or
b00 and the price you pay B is independent of your bid. This gives the matrix of payoffs
below. Again you can see that the top row is weakly dominated by the bottom row.
B > vi B ∈ (b00 , vi , ) B < b00
Bid b00 Lose = 0 Lose = 0 Win = vi − B > 0
Bid vi Lose = 0 Win = vi − B > 0 Win = vi − B > 0
Hence we can conclude: it is a weakly dominating strategy to bid truthfully in a second
price auction.
5. DYNAMIC OLIGOPOLY
MARTIN CRIPPS
1. Introduction
In the models of oligopoly we studied in the previous section we imagined that the firms
were moving simultaneously, either choosing quantities (Cournot) or prices (Bertrand) at
exactly the same time. Thus the firms were playing a game in strategic form. The right way
to study games in strategic form is to use the tools of Nash equilibrium and dominance.
So this is how we modelled competition in these games.
Now we are going to study games where there is an order to the moves. In the first
two sections we will study competition in the case where one firm moves first and then
the other firm moves. In the final section the firms will move at the same time but there
are many periods where they can make different moves. Thus we are considering games
in extensive form. In games like these we know Nash equilibrium is inadequate, as it may
have non-credible threats. So the way we model competition now will be to use either
backwards induction or subgame perfection.
2. Stackelberg Competition
Stackelberg criticised Cournot’s way of modelling quantity competition, because he be-
lieved industries always had one firm that got to dominate the competition in a particular
way. In particular he thought there was a specific order in which the firms got to make
their quantity choices, there was a leader and then other firms followed, and that this order
affected the quantity choices the firms in the industry made.
This can obviously be quite a complicated situation to describe as a game tree, if there
are many firms in an industry. But in its simplest form, with only two firms, we could
imagine that first firm 1 (the leader) chooses its quantity, then firm 2 (the follower) chooses
its quantity, then the price is determined, and finally the firms’ profits are realised. Here
is the simplest possible version of this in game-tree form.
Firm%1%
q1=3% q1=4%
Firm%2% Firm%2%
1
2 MARTIN CRIPPS
In this game Firm 1 chooses a quantity q1 = 3 or q1 = 4, then firm 2 sees the quantity
chosen by the leader and chooses a quantity q2 = 3 or q2 = 4 themselves. How does this
change the outcome of competition? Well it depends on what kind of equilibrium we look
for. Stackelberg said we should solve this game using backwards induction (he introduced
this idea long before game theorists took it up). Here is the backwards-induction solution
to the game.
Firm%1%
q1=3% q1=4%
Firm%2% Firm%2%
First look at what firm 2 (the follower) does. It takes Firm 1’s output as given and
chooses the profit-maximizing output given Firm 1’s choice. Firm 2 is choosing its best
response to every output choice of Firm 1. Thus, in the terminology of reaction functions,
we would say that whatever Firm 1 does Firm 2 is always making a choice on its reaction
function.
The consequence of this structure is the leader has an advantage, because it can produce
a lot and force the follower to cut back its own output. Thus, in contrast to the Cournot
equilibrium the firms now behave in different ways although they have the same costs and
demand.
Now we will generalise these ideas to the firms we studied in Cournot competition.
so many Q1 ’s to consider. How can we do backwards induction without the picture? Here’s
how. . .
2.1.1. Follower’s Behaviour. If we think about the previous example, what the follower
did was to choose the best Q2 given the observed Q1 . This is exactly what the reaction
function of player 2 does. It describes the best possible Q2 for each Q1 . So, the last stage
of the backwards induction is precisely described by the function.
1
Q2 = 12 − Q1 Final Stage of Backwards Induction
2
2.1.2. Leader’s Behaviour. The leader knows how the follower will respond to any Q1 they
choose. If they choose Q1 = 3 the follower will choose Q2 = 12 − 3/2 = 10.5, because this
is their optimal response as determined by their reaction function. How should the leader
choose Q1 then? Well it should choose Q1 to maximise its profit taking into account the
fact that Q2 is constrained to equal 12 − Q1 /2. That is, the leader solves the problem:
1
max(30 − Q1 − Q2 )Q1 − 6Q1 , subject to Q2 = 12 − Q1 .
Q1 2
This constrained optimization can be solved by substitution:
1
max 30 − Q1 − 12 − Q1 Q1 − 6Q1 .
Q1 2
Simplifying this gives:
1 1
max 30 − Q1 − 12 − Q1 Q1 − 6Q1 = max 12Q1 − Q21 .
Q1 2 Q1 2
Differentiating and setting equal to zero gives:
Q1 = 12.
We can substitute this back into the reaction function of Firm 2 to get Q2 = 6.
Let us compare this situation with that when the firms move simultaneously (under
Cournot competition). If there is Cournot competition the firms will both produce output
8 (Q1 = Q2 = 8), so the total supply is 16 and the price is 14. If there is Stackelberg
competition the outputs are Q1 = 12 and Q2 = 6, so the total supply is 18 and the price
is 12. This is quite generally true. Under Stackelberg competition the leader produces
more the follower produces less and the leader makes more profit than under Cournot
competition. Thus being a quantity leader confers considerable benefits on the leader and
harms the followers.
3. Price Leadership
Just as we can think of situations where the firms set quantities in a definite order so we
can also imagine situations where firms set prices in an order too. Let us also imagine that
Firm 1 is the price leader and Firm 2 is the price follower. Here the effects of the order are
entirely different—being a leader in pricing harms the leader and benefits the follower. In
its starkest form suppose the firms are producing identical products, then if Firm 1 goes
4 MARTIN CRIPPS
first and sets price, then Firm 2 can undercut this price and take all the customers. Here’s
a simple example of this.
Firm%1%
P1=3% P1=4%
Firm%2% Firm%2%
Here the second mover has and advantage and this is transparent when we consider the
backwards-induction solution.
Firm%1%
P1=3% P1=4%
Firm%2% Firm%2%
Now let us do price leadership in the case when the firms are not producing identical
products. Recall our model of price competition with differentiated products where as the
goods are not identical, demand does not jump to zero for the highest price firm
Firm1 Demand Q1 = 12 − 2P1 + P2 ,
Firm2 Demand Q2 = 12 + P1 − 2P2 .
This gave rise to the (Price) reaction functions
1
P1 = 3 + P2 ,
4
1
P2 = 3 + P1 .
4
and a Nash equilibrium in the simultaneous move game where P1 = P2 = 4.
3.0.3. Follower’s Behaviour. Just as in the quantity-setting game the price-follower (Firm
2) will choose their price to be an optimal response to what the leader does. Thus Firm 2
will see P1 and then choose
1
P2 = 3 + P1 .
4
5. DYNAMIC OLIGOPOLY 5
3.0.4. Leader’s Behaviour. The leader wishes to maximise their profit by choosing the right
P1 . In the previous section we noted that the leader’s profit was equal to
π1 = P1 Q1 − 20 = 12P1 − 2P12 + P1 P2 − 20.
However,the leader is constrained because whatever P1 they chose will lead to the choice
of a P2 = 3 + 0.25P1 by the follower. Thus the leader solves the problem:
1
max 12P1 − 2P12 + P1 P2 − 20, subject to: P2 = 3 + P1 .
P1 4
This can be solved by substitution:
1
max 12P1 − 2P12 + P1 3 + P1 − 20 = 15P1 − 1.75P12 − 20.
P1 4
Differentiating this and setting this equal to zero to find the maximum gives
2
0 = 15 − 3.5P1 or P1 = 30/7 = 4 + .
7
Thus the price leader chooses to set a higher price than 4 (the equilibrium in the simulta-
neous move game). The price follower sets the price
1 30 1
P2 = 3 + =4+
4 7 14
and undercuts the leader. They do well at this lower price.Their demand is higher than at
4 because the leader has raised their price and they the follower are undercutting them.
To add up the profits over many periods we will use the rate of interest r to calculate
the firm’s discounted present value of the profit it earns every period. For example, if the
firm earns 7 units of profit every period of the game we will say its payoff from playing the
entire repeated game is
7 7 7
=7+ + 2
+ + ...
(1 + r) (1 + r) (1 + r)3
(Recall that r > 0 is an interest rate.) We can use what we know about geometric series
to work out this sum.
7 7 7
Payoff = 7 + + + + ...
(1 + r) (1 + r)2 (1 + r)3
1 1 1
=7 1+ + + + ...
(1 + r) (1 + r)2 (1 + r)3
∞
X 1
=7
(1 + r)t
t=0
1
=7 1
1− (1+r)
1+r
=7
r
Tis repeated game has many equilibria. Now we will give an example of one equilibrium
which is particularly beneficial to the firms. When the play out the equilibrium the firms
coordinate on the high price and only threaten to play a low price if one player cheats on
this. We will assume they both play the strategy:
“Set a high price if my opponent has set the high price in every past period,
but play the low price in all other cases.”
Such a strategy promotes cooperation by making something very bad happen if a player
cheats on the cooperation. (The very bad thing is permanently low prices.) These strategies
are usually called “Grim Trigger” strategies by game theorists.
To see why it is an equilibrium first notice that if I stick to playing the high price (and
my opponent does) the previous calculation tells me I will have profits
10(1 + r)
Payoff of Not Cheating = .
r
If I cheat on high prices I get 12 for one period and then 7 forever, so in total I get
7 7 7
Payoff of Cheating = 12 + + 2
+ + ...
(1 + r) (1 + r) (1 + r)3
∞
7 X 1
= 12 +
(1 + r) (1 + r)t
t=0
7
= 12 +
r
5. DYNAMIC OLIGOPOLY 7
ECON2001 Microeconomics
Lecture Notes
Term 1: Production
Ian Preston
Technology
Suppose a firm produces a single output q with inputs z. The constraints
imposed by technology can be represented using a production function f(z) so
that it is feasible to produce q using z if q ≤ f (z). Production is technically
efficient if q = f (z) so the greatest possible output is being produced given the
inputs.
More general than this is a formulation representing a firm’s activities by
a production plan, production vector or net output vector, y, in which outputs
and inputs are not distinguished. Negative quantities in a particular production
plan correspond to goods consumed (inputs) and positive quantities to goods
produced (outputs). This allows for multiple outputs and also for certain goods
to be either inputs or outputs depending on the production plan. In this more
general setting we can represent technological feasibility with what is called a
transformation function, F (y), with feasible plans satisfying F (y) ≤ 0. For
the single output case, y = (q, −z) and F (q, −z) = q − f (z). The transforma-
tion frontier is the set of technically efficient production plans, which is to say
satisfying F (y) = 0.
Properties of production
• Possibility of shutdown: Shutdown is usually assumed possible which
means that 0 ≤ f (0) or, in the more general case, F (0) ≤ 0. If no output
can be produced without using some inputs then these are equalities.
• Monotonicity: If production is technically efficient then net output of
one good can be increased only by decreasing net output of another. Out-
put can only be increased by increasing some input. To produce the
same output after decreasing one input requires increasing another. If
there is only one output then the production functions must be increasing
∂f /∂zi ≥ 0. In the more general formulation ∂F/∂yi ≥ 0.
∂F ∂F dyj
+ = 0
∂yi ∂yj dyi F (y)=0
dyj ∂F /∂yi
⇒ M RTij = = −
dyi F (y)=0 ∂F /∂yj
dq ∂f
M Pi = =
dzi q=f (z) ∂zi
(It may seem pedantic to note it but it may help to avoid confusion in
later uses of the idea to note that since inputs are minus net outputs this
is minus the marginal rate of transformation between net outputs in this
case.)
• The rate at which any input has to be increased as we decrease another
holding output constant is known as the marginal rate of technical substi-
tution between the two and is equal to the slope of the isoquant
dzj ∂f /∂zi
M RT Sij = − =
dzi q=f (z) ∂f /∂zj
Homotheticity
Production is homothetic if, whenever input vectors z A and z B produce the
same output, f (z A ) = f (z B ), then scaling those input vectors up or down by
the same factor will also give input vectors which produce the same output.
In other words, for any λ > 0, f (λz A ) = f (λz B ). This means that isoquants
corresponding to higher levels of output are just magnified versions of isoquants
producing lower levels of output. It also means that marginal rates of technical
substitution are constant along rays through the origin.
Production is said to be homogeneous of degree α if f (λz) = λα f (z). Ho-
mogeneous production functions always have the property of homotheticity.
Returns to scale
Returns to scale are concerned with the feasibility of scaling up and down
production plans. (Remember that a production plan includes a specification
of outputs as well as inputs so we are now talking about scaling up inputs and
output.)
• There are decreasing returns to scale (DRS) if scaling up the input vector
results in a less than proportionate increase in output. Thus λf (z) >
f (λz) for λ > 1. It is therefore possible to scale production plans down
but not up.
If production is homogeneous then there are decreasing returns to scale if
α < 1.
If there is only one output and one input then concavity of the production
function, f 00 < 0, implies decreasing returns to scale given f (0) = 0.
• There are increasing returns to scale (IRS) if scaling up the input vector
results in a more than proportionate increase in output. Thus λf (z) <
f (λz) for λ > 1. It is therefore possible to scale production plans up but
not down.
If production is homogeneous then there are decreasing returns to scale if
α > 1.
If there is only one output and one input then convexity of the production
function, f 00 > 0, implies increasing returns to scale given f (0) = 0.
• There are constant returns to scale (CRS) if scaling up the input vector
results in exactly the same scaling up of production. Thus λf (z) = f (λz)
for any λ. This is equivalent to homogeneity of degree one of the produc-
tion function. If there is also only one input then production is linear,
q = az for some a.
Some examples
• Perfect substitutes, CRS q = f (z1 , z2 ) = az1 + bz2 : MP of z1 is a
and MP of z2 is b and both are constant. The MRTS is the ratio of the
marginal products a/b and is constant. Isoquants are therefore parallel
straight lines. The elasticity of substitution is infinite. Production is
homothetic and there are constant returns to scale.
• Perfect substitutes, general q = f (z1 , z2 ) = G(az1 + bz2 ): The MRTS
is still constant at a/b, isoquants are therefore still parallel straight lines
and the elasticity of substitution is still infinite. Production is also homo-
thetic as the MRTS is constant along rays (because it is constant every-
where). Returns to scale depend however on the function G.
• Perfect complements, CRS q = f (z1 , z2 ) = min[az1 , bz2 ]: MP of z1
is a if az1 < bz2 and zero otherwise; MP of z2 is b if az1 > bz2 and zero
otherwise. Isoquants are L-shaped with the kinks lying on a ray through
the origin of slope a/b. Production is homothetic and there are constant
returns to scale.
• Cobb-Douglas: q = f (z1 , z2 ) = Az1a z2b : MP of z1 is aq/z1 and MP of
z2 is bq/z2 . The MRTS is therefore az2 /bz1 which is diminishing so input
requirement sets are convex. Since ln(z1 /z2 ) = ln(a/b) − ln M RT S the
elasticity of substitution is constant at 1. The MRTS depends only on
the input ratio z1 /z2 so production is homothetic. Since f (λz1 , λz2 ) =
α β
A (λz1 ) (λz2 ) = λa+b f (z1 , z2 ), the production function is homogeneous
of degree a + b and shows IRS if a + b > 1, DRS if a + b < 1 and CRS if
a + b = 1.
b
• Constant elasticity of substitution: q = f (z1 , z2 ) = [Az1a + Bz2a ]
with a < 1: MP of z1 is abAq 1−1/b z1a−1 and MP of z2 is abBq 1−1/b z2a−1 .
a−1
The MRTS is therefore (A/B) (z1 /z2 ) which is diminishing so input
requirement sets are convex. Since ln(z1 /z2 ) = [ln(a/b) + ln M RT S] /(a−
1) the elasticity of substitution is constant at 1/(1 − a). The MRTS
depends only on the input ratio z1 /z2 so production is homothetic. Since
f (λz1 , λz2 ) == λab f (z1 , z2 ), the production function is homogeneous of
degree ab and shows IRS if ab > 1, DRS if ab < 1 and CRS if ab = 1.
√
• A non-homothetic
√ example: q = f (z1 , z2 ) = z 1 + √ z2 The MP of z1
is 1/2 z 1 and the MP of z2 is 1. MRTS is therefore 1/2 z 1 which is not
constant along rays so production is not homothetic. Input requirement
sets are convex but the elasticity of √ substitution
√ is not constant. There is
DRS if z1 > 0since f (λz1 , λz2 ) = λ z 1 + λz2 < λf (z1 , z2 ) if λ > 1 (and
CRS if z1 = 0).
• A multiple product example: Suppose a firm uses a single input z
which it can allocate to production
√ of either √
of two goods. If it uses ζ
in producing good 1 then q1 = ζ and q2 = z − ζ. Thus q12 + q22 = z.
Profit maximisation
Suppose there is a single output priced at p, input prices are w and the firm
regards these prices as given. This would be sensible for a small firm under
competitive conditions. Firm profits are pf (z) − w0 z and we assume that the
firm makes production decisions so as to maximise its profits.
First order conditions, assuming all inputs are used in positive but finite
quantities at the optimum, require
∂f
p − wi = 0
∂zi
for each input, or equivalently
M Pi = wi /p.
The firm uses the input until its marginal product equals its input price divided
by the price of the firm’s output. In other words, the price ratio between input
and output is equated to (minus) the marginal rate of transformation between
output and input.
Second order conditions include the requirement that ∂ 2 f /∂zi2 ≤ 0 so that
output is a concave function of each input.
If we take the conditions relating to any two inputs then we can divide
through to eliminate the product price
∂f /∂zi
= wi /wj
∂f /∂zj
pA q A − wA z A ≥ pA q B − wA z B
pB q A − w B z A ≤ pB q B − w B z B
pA ∆q − wA ∆z ≥ 0 pB ∆q − wB ∆z ≤ 0
and therefore
∆p∆q − ∆w∆z ≥ 0.
Thus if output price goes up (∆p > 0) holding input prices constant
(∆w = 0) then output cannot fall (∆q ≥ 0). Likewise if one input price
rises (∆wi > 0) while output price stays the same (∆p = 0) as do all
other input prices (∆wj = 0, j 6= i) then demand for that input cannot
rise (∆zi ≤ 0).
Hence S(p, w) is increasing in p and Di (p, w) is decreasing in wi .
• Increasing the price of output cannot decrease profit and increasing the
price of an input cannot increase profit.
Hotelling’s Lemma
If we take π(p.w) and differentiate with respect to output price p then
∂π X ∂f
∂Dj (p, w)
= f (D(p, w)) + p − wj
∂p j
∂zj ∂p
= f (D(p, w)) = S(p, w)
using the first order condition that says p∂f /∂zj − wj = 0. Hence the derivative
is the supply function.
Similarly, if we differentiate with respect to an input price wi then
∂π X ∂f
∂Dj (p, w)
= −Di (p, w) + p − wj
∂wi j
∂zj ∂wi
= −Di (p, w)
Symmetry
Given these relationships
Cost minimisation
For firms producing a single output, it is useful to model firm decision making
as a two-stage process with inputs chosen to minimise costs given output q and
output q then chosen to maximise profit given what that implies for how costs
vary with q. An implication of profit maximisation is that the chosen output
level is always produced at least cost.
This is a useful perspective because firms may not be price takers in output
markets but nonetheless still act as input-price-taking cost minimisers in input
markets.
Consider, then, a firm choosing inputs z to solve
w0 z + µ [q − f (z)]
µ∂f /∂zi − wi = 0
and therefore imply equality between the marginal rate of technical substitu-
tion and input price ratio, M RT Sij = wi /wj exactly as derived in the profit
maximisation problem above.
We can visualise the optimum choice as occuring at a tangency between an
isoquant and an isocost, at a point where the slope of the isocost equals the
ratio of input prices.
wA z A ≤ wA z B wB z A ≥ wB z B
∆w∆z ≤ 0.
If only one input price changes so that ∆wi > 0 and ∆wj = 0 for i 6= j,
then ∆zi ≤ 0. Conditional input demands must therefore be decreasing
in own price.
Cost function
The function giving minimum cost given q and z is known as the cost func-
tion, C(q, w) = w0 H(q, w). It is
• increasing in the price of every input which is used
Shephard’s Lemma
If we differentiate the cost function with respect to one of the input prices
then we get a relationship linking the derivative to the conditional input demand.
∂C X ∂Hj (p, w)
= Hi (q, w) + wj
∂wi j
∂wi
X ∂f ∂Hj (p, w)
= Hi (q, w) + µ
j
∂zj ∂wi
= Hi (q, w)
where the unconditional demands are recognised as the conditional input de-
mands at profit maximising supply.
Differentiating, we see that
∂Di (p, w) ∂Hi (S(p, w), w) ∂Hi (S(p, w), w) ∂S(p, w)
= + ·
∂wi ∂wi ∂q ∂wi
so the unconditional effect of increased input price includes both an effect at
fixed output ∂Hi /∂wi and an effect coming through adjustment to chosen output
level ∂Hi /∂q · ∂S/∂wi .
C(q, w) = qκ(w).
If there are increasing returns to scale (IRS) on the other hand then it is possible
to scale production up but not down and multiplying output by λ > 1 will
multiply costs by less than λ
The first order condition for this problem requires that price be equated to
marginal cost, p = ∂C/∂q. The second order condition requires ∂ 2 C/∂q 2 ≥ 0.
Because shutdown has been assumed possible, firms cannot make negative
profits. The condition that price exceed average cost p ≥ C/q can therefore be
imposed on firm decisions. Given that p = ∂C/∂q, this implies ∂C/∂q ≥ C/q
so that it is only output ranges over which average costs are rising that will be
chosen. The competitive firm’s supply curve is therefore that segment of the
marginal cost curve lying above the average cost curve.
H1 (q, w) = q 2 /a H2 (q, w) = q 2 /b
MARTIN CRIPPS
1. Introduction
We now study models of markets where there is uncertainty about the type of the buyers
or the type of the sellers. If a buyer knows what type she is but the seller doesn’t we are in
a situation of asymmetric information. (One person in the game/market knows more than
the other.) In models of trading that economists study, asymmetric information gives rise
to a phenomenon termed “adverse selection”. We will see that uncertainty (or asymmetric
information) has profound effects on markets. In particular uncertainty makes optimization
difficult, harms efficiency and introduces a role for signalling.
are bad, the buyers are very optimistic and believe all cars are good, or some intermediate
belief. We will assume that buyers are not completely ignorant and are not dogmatic, but
that their beliefs reflect the aggregate mix of good and bad cars being traded. That is,
in an equilibrium the buyers know the general shares of good and bad cars for sale at the
current price.
Now let us start analysing the model. First we plot the two supply curves for good and
bad used cars. The
Thesupply
supplyof good used cars
of good lies(green)
cars above theis
supply
aboveof bad
theones (less are
supply of bad cars (black),
supplied at each price).
because sellers of a good car require a higher price to be willing to part with it.
P"
Q"
If these two supply curves are summed, we get the total supply of cars at each price (red).
The total supply of good and bad cars is8" the horizontal sum of good and
Observe that bad
at low prices
cars supply. only the bad used cars are offered and as the price rises the
share of good used cars increases.
P"
Q"
Now we draw the demand for used cars9" from the buyers. This will obviously depend
upon the buyers’ beliefs about the quality of cars being offered for sale. This figure has 3
demand curves. The lowest (light pink) is the demand curve if the buyers believe all used
cars are bad. The highest demand curve is the demand if the buyers think all used cars
are good. And the intermediate demand is when the buyers think the mix is 50:50. We
cannot draw the demands for all possible beliefs, but we assume that they will smoothly
adjust downwards as the buyers become more pessimistic about quality.
If buyers think all cars are good then demand is maximized.!
6. ADVERSE SELECTION 3
P#
Q#
Now we combine the demands and supply. Suppose first that the buyers were optimistic
12#
and believed that all the cars sold were good. Then demand equals supply at the rightmost
blue dot in the figure below. At this point it is clear that there are not all good cars being
sold but a mix of good and bad. Thus the buyers should become more pessimistic and the
demand should shift down. As the demand shifts down, less and less good cars are sold at
equilibrium. In some In the limit
cases only will
demand bad continue
cars are sold and buyers
to shift down know this.!bad cars a sold
until only
and buyers believe only bad cars are sold.
P’’’#
Notice that sellers of good cars would like to be able to convince the buyers that they
are selling a good car, but they cannot because anything
15# they say could also be said by
low-quality sellers. That is, there is no credible way of distinguishing your good car from
someone else’s bad car. As a result low quality squeezes out high quality. This is called
Adverse Selection. One interpretation of what is going on here is that buyers lower prices
to protect themselves against the risk of buying a bad used car. But by lowering prices
they encourage fewer good cars to be sold.
Adverse selection is important in financial markets, because sellers may be selling a stock
because the stock is bad and buyers may be buying because the stock is good. This leads
to Bid-Ask spreads: Sellers get offered lower prices to protect the market from those people
who are selling because the news is bad. Buyers get offered higher prices to protect the
market from those people who are buying because the stock is good.
4 MARTIN CRIPPS
2.1. Cursed Equilibrium. In the Lemons model, the buyers face a problem of adverse
selection. Another word for this problem is the winner’s curse. That is, when a buyer
succeeds in getting a car they may come to regret their success because the car is worse
than they thought it would be. This is a feature of many markets where there is less
than perfect information and competition among buyers—the winners curse is particularly
problematic in auctions.
In the above model the buyers are very smart and they make the correct guess about
the proportion of good and bad cars in the market, so they will not be disappointed on
average. One extension to the lemons model would be to allow some of the buyers to have
too optimistic beliefs about the quality of cars. That is, a share of the buyers are irrational
and believe the cars are good while the rest of the buyers correctly perceive the share of
good and bad cars. This allows us to incorporate irrationality into this world pretty easily.
This is studied in papers on cursed equilibrium
3.1. The Model. The model will derive a demand for insurance at each price. The price
will also determine the type of customers who demand insurance. To find an equilibrium
in this market we will then use the condition that perfectly competitive firms must make
zero profit. It is this zero-profit condition that will ultimately determine the equilibrium
in this market. The first step is to describe the buyers
3.1.1. The Buyers of Insurance. There is an infinity of buyers of insurance and the buyers
come in many different types. Each buyer’s type is described by π, which is their probability
of having an accident. We suppose their is a continuum of buyers. And the aggregate
distribution of buyers 0 ≤ π ≤ 1 is a uniform distribution on the interval [0, 1].
B.#Market#Failure#and#Asymmetric#Info#8
Assumption#:#Suppose#that#there#is#a#continuum#of#
customers#with#types#that#are#distributed#uniformly#on#
6. ADVERSE SELECTION 5
the#interval#between#0#and#1.8
π#:=Pr(This#consumer#has#an#accident)8
8
88 1!
8
8
π!
0! 1!
So for example if we found that buyers with riskiness above π = 3/4 brought insurance,
we would have a demand for insurance that was equal to 1/4.
Now we describe the buyers’ utility for insurance. We will assume that each buyer has
a wealth w > 0, but if they have an accident they will incur some costs L. So as a result of
their accident their wealth will fall to w − L. We will also assume each buyer has a utility
of wealth U (w) = w1/2 .
Now let us consider the expected utility of a buyer that has a risk off accident π. With
probability π they will have an accident and end up with utility (w−L)1/2 . With probability
1 − π they will not have an accident and end up with utility w1/2 . Thus the buyer has
expected utility
This inequality is very important to the subsequent analysis. First, notice that we do have
adverse selection, because this inequality says that only types with high riskiness will buy
insurance. That is, all the types with π satisfying:
1/2
1 − 1 − Pw
h(P ) = <π≤1
L 1/2
1− 1− w
strictly prefer to buy insurance. We will write the function on the left here as h(P ) so
all those people with h(P ) ≤ π ≤ 1 will buy insurance. How does the number of people
change who buy insurance when the price adjusts. Well it is not to hard to see that as P
increases so does h(P ) so fewer people buy insurance. It is also easy to see
1 − (1 − 0)1/2
h(0) = = 0,
L 1/2
1− 1− w
L 1/2
1− 1− w
h(L) = = 1.
L 1/2
1− 1− w
B.#Market#Failure#and#Asymmetric#Info#8
So if insurance were free everyone would buy it and if insurance cost as much as the loss
you were going to make no-one would buy it.
Once we Supply#is#determined#by#average#riskiness#of#the#buying#
know who is going to buy insurance we can work out the demand for insurance
at each price.customers.8
Below we have a picture of the mass of consumers who choose to buy
8
insurance at the price P .
8 h(P)!
88
Not!Buy! Buy!
8 1!
8
0! π!
1!
The total area in this figure is the size of the demand for insurance at the price P . This
area is equal to 1 × (1 − h(P )). So we now can write down the demand for insurance at
the price P
1/2 L 1/2
1 − Pw
− 1− w
Demand(P ) = 1 − h(P ) = .
L 1/2
1− 1− w
3.1.2. The Sellers of Insurance. There is one aspect of the buyers’ behaviour that will
particularly matter to the sellers and that is how risky are the customers when the current
price is P .
First let us investigate how sellers behave if they know the riskiness of their customer.
Consider a customer with known risk π. A seller will have a cost equal to zero for the
6. ADVERSE SELECTION 7
customer if they don’t have an accident and a cost L for the customer if they do. This
makes the expected cost of the customer
Cost of Customer(π) = πL + (1 − π)0 = πL.
As the seller charges the price P , the profit from a customer of known riskiness is P − πL.
Thus, as long as P ≥ πL the firm will make a profit from selling to this customer if it
knows their riskiness.
Now suppose theB.#Market#Failure#and#Asymmetric#Info#8
price is P and the firm does know know the riskiness of their customers.
Of course, at an equilibrium they do know that only customers with riskiness above h(P )
will buy insurance. So what is the average riskiness of the customers that do buy insurance
Firms#costs#are#determined#by#average#riskiness#of#a#
at the price P ?customer.8
In our particular example this can be easily seen from the following picture.
8 Average!Riskiness!=0.5[1+h(P)]!
8
88
Not!Buy! Buy!
8 1!
8
π!
0! h(P)! 1!
The halfway point between h(P ) and 1 is (1 + h(P ))/2, so this is the average of the
customers in the blue region.1
Now we have the average riskiness of the customers that buy when the price is P , we
can work out the firm’s costs from selling insurance in this case. It is
Expected Costs = L(1 + h(P ))/2.
So now we can work out the firms expected profit from selling insurance as P − L(1 +
h(P ))/2.
We know that firms will enter the industry as long as the make strictly positive profit
from selling insurance so we could now write the supply curve in this market as:
(
∞ : P ≥ L(1 + h(P ))/2
Supply =
0 : P < L(1 + h(P ))/2
1BUT, this is the average riskiness here because the distribution is flat. In general it will not be possible
to just draw a line like this. The correct way to calculate the average riskiness for an arbitrary distribution
f (π) would be the following conditional expectation
R1
h(P )
πf (π)dπ
Average Riskiness of Buyers in[h(P ), 1] = R 1
h(P )
f (π)dπ
where f (π) = 1 is the density of the distribution of π. You can check this here by substituting in f (π) = 1
and doing the calculation.
8 MARTIN CRIPPS
This gives the supply curve as a horizontal line where P = L(1 + h(P ))/2. Where this
crosses the demand curve above gives the equilibrium in this market.
3.1.3. Equilibrium in a Perfectly Competitive Economy. To find the equilibrium price we
must find where supply equals demand. As supply will be zero or infinite if firms are not
making zero profits, to find equilibrium we find the place where firms make exactly zero
profit. That is, we must solve the equation
B.#Market#Failure#and#Asymmetric#Information8
P = L(1 + h(P ))/2.
Here is a picture ofJoining#dots#and#using#the#fact#that#h(P)#is#increasing8
the two sides of this equation
8 8 8#L[1+h(P)]/2#=#P8
P=revenue"
L"
L[1+h(0)]/2"
Expected"Costs"
B.#Market#Failure#and#Asymmetric#Information8
P"
L"
From this alone we can see that there must be an equilibrium where P = L and no-one
But#there#could#be#circumstances#in#which#some#(but#not#
buys insurance. Butall)#people#get#insurance#(L[1+h(P)]/2#=#P).8
maybe there are more equilibria and the picture looks like
P=revenue"
L"
L[1+h(0)]/2"
Expected"Costs"
P"
L"
It is quite hard to determine this. One way is to look at the slope of the function L(1 +
h(P ))/2 where P = L. this is because if the slope of this function is more than 1 when
P = L it means the picture looks like the above. (This has been set as an exercise.)
3.1.4. Equilibrium with Monopoly. We could also consider the problem of a monopolist
selling insurance. Then, if the monopolist set the price P they would sell to 1 − h(P )
customers. Each would pay the price P and on average the customers would cost the firm
L(1 + h(P ))/2. Thus the monopolist would make the profit
(1 − h(P ))(P − L(1 + h(P ))/2).
The monopolist would then choose P to maximise this expression.
6. ADVERSE SELECTION 9
#7
ψ`(x)"=""Demand"Curve"for"good"x"
P"
x"
x*(p)"
So the customer will buy the deal x∗ at the price p and fee F if ψ(x∗ ) − px∗ ≥ F and will
not buy if ψ(x∗ ) − px∗ < F . That is, the customer either buys the quantity where marginal
benefit equals marginal cost (when the value it receives is greater than the fixed cost) or
nothing. Another way of saying this is that if the fixed fee is less than the total value the
C.#Two'Part#Tariffs#and#Adverse#Selection7
customer obtains then she is prepared to buy the good. The picture shows the optimal
fixed fee where ψ(x∗ ) − px∗ = F .
#7
The"fixed"fee"is"the"
ψ`(x)" area"below"the"
demand"curve"and"
above"the"per"unit"
F" sales."
P"
px*$
x"
x*(p)"
Now we can think about what is the price-fee combination that maximises the firm’s
profit. We suppose that the monopolist has a cost per unit of c > 0. As usual one of the
6. ADVERSE SELECTION 11
P"
px*$
x"
x*(p)"
Adverse selection is now present in this model, because the monopolist would like to offer
to sell the good at marginal cost and take all the value. (But if it did this it would have to
offer a lower fixed fee to the low-value customers and a higher fixed fee to the high-value
12 MARTIN CRIPPS
customers.) This is impossible as all the high-value customers would then pretend to be
low-value customers.
The screening solution to the adverse selection problem. Now we will describe the menu
of deals the monopolist. The first deal (xa , Fa ) is a quantity and a payment (this payment
covers both the fixed fee and the per unit price). The monopolist would like this deal to
be bought by the high-value customers. The second deal (xb , Fb ) is a second quantity and
payment, which the monopolist would like to be bought by the low-value customers. In
total the monopolist will offer a menu of deals ((xa , Fa ); (xb , Fb )). Suppose the monopolist
were successful in getting the two types of customers to buy the appropriate deal, then it
would make the profit.
Profits = φ(Fa − cxa ) + (1 − φ)(Fb − cxb ).
4.2.1. Getting the Customers to Buy: Individual Rationality Constraints. The first thing
to notice is that the high-value customers are not willing to buy any deal. If the price was
very high they would rather not buy anything. If they are going to buy the deal (xa , Fa )
then they need to get positive from it. That is,
High-Value’s Utility from A-Deal ≥ Utility from not buying
αψ(xa ) − Fa + y ≥ y.
Or αψ(xa ) ≥ Fa . Similarly if the monopolist is going to get the low value types to buy the
B-Deal, then
Low-Value’s Utility from B-Deal ≥ Utility from not buying
βψ(xb ) − Fb + y ≥ y.
Or βψ(xb ) ≥ Fb . The result of these calculations are two constraints on the kinds of deals
the monopolist can choose to maximise its profit
αψ(xa ) ≥ Fa βψ(xb ) ≥ Fb .
These are called Individual Rationality Constraints which kind of means you cannot force
customers to buy things they don’t want to buy. They are sometimes also called partici-
pation constraints.
4.2.2. Getting the Customers to Buy the Right Deal: Incentive Compatibility Constraints.
The second thing the monopolist needs to worry about is that the high-value customers
actually choose to buy the low-value customers’ deal.(xb , Fb ) was offering the good at a
very low price the high-value types would rather buy this than (xa , Fa ). If they are going
to buy the deal (xa , Fa ) then they need to get higher utility from it than from (xb , Fb ).
That is,
High-Value’s Utility from A-Deal ≥ High-Value’s Utility from B-Deal
αψ(xa ) − Fa + y ≥ αψ(xb ) − Fb + y.
6. ADVERSE SELECTION 13
Or αψ(xa ) − Fa ≥ αψ(xb ) − Fb . Similarly if the monopolist is going to get the low value
types to buy the B-Deal and not the A-Deal
Low-Value’s Utility from B-Deal ≥ Low-Value’s Utility from A-Deal
βψ(xb ) − Fb + y ≥ βψ(xa ) − Fa + y.
Or βψ(xb ) − Fb ≥ βψ(xa ) − Fa . The result of these calculations are two constraints on the
kinds of deals the monopolist can choose to maximise its profit
αψ(xa ) − Fa ≥ αψ(xb ) − Fb βψ(xb ) − Fb ≥ βψ(xa ) − Fa .
These are called Incentive Compatibility Constraints which kind of means you are giving
the customers the incentive to buy the things you want them to buy.
4.2.3. Reducing the Number of Constraints. Now we have found four different constraints
on the monopolist’s choice of a menu ((xa , Fa ); (xb , Fb )). Here is the list
(1) αψ(xa ) − Fa ≥ 0
(2) βψ(xb ) − Fb ≥ 0
(3) αψ(xa ) − Fa ≥ αψ(xb ) − Fb
(4) βψ(xb ) − Fb ≥ βψ(xa ) − Fa
Consider the individual rationality constraint for the high value types, (1), this is ensured
because the A’s could pretend to be low-value types and they can get positive value when
they do this. Here is the argument. First the high values don’t want the low-value deal,
this is the constraint (3)
(3)
z}|{
αψ(xa ) − Fa ≥ αψ(xb ) − Fb
Then we notice that α > β, so αψ(xb ) ≥ βψ(xb ) and
(3) α≥β
z}|{ z}|{
αψ(xa ) − Fa ≥ αψ(xb ) − Fb ≥ βψ(xb ) − Fb .
Finally we use the fact that this last expression is the low type’s utility from buying the
B-deal, which must be positive.
(3) α≥β (2)
z}|{ z}|{ z}|{
αψ(xa ) − Fa ≥ αψ(xb ) − Fb ≥ βψ(xb ) − Fb ≥ 0.
Thus by using (3) and α ≥ β and (2) we can conclude that αψ(xa ) − Fa ≥ 0 which is (1).
Thus this condition is implied by the remaining three constraints and we only need them.
Here is the list:
(5) βψ(xb ) − Fb ≥ 0,
(6) αψ(xa ) − Fa ≥ αψ(xb ) − Fb ,
(7) βψ(xb ) − Fb ≥ βψ(xa ) − Fa .
14 MARTIN CRIPPS
Now think about adding a small amount f to the payments Fa and Fb . If you did this the
new constraints would be:
βψ(xb ) − Fb − f ≥ 0,
αψ(xa ) − Fa − f ≥ αψ(xb ) − Fb − f,
βψ(xb ) − Fb − f ≥ βψ(xa ) − Fa − f.
This change would increase what the monopolist earned from all its customers. And so
would have to be good for the monopolist. The change would not be a problem for the
two last constraints, because the f affects both sides of these equally. However, it might
be a problem for the first constraint, because as f gets bigger eventually it would become
negative. Thus we can deduce that the first constraint has to be an equality when the
monopolist is maximising its profit.
βψ(xb ) − Fb = 0
+
z}|{
αψ(xa ) − Fa ≥ αψ(xb ) − Fb
+
z}|{
βψ(xb ) − Fb ≥ βψ(xa ) − Fa
Now the first constraint pins down Fb , but the monopolist would still like to make Fa as big
as possible. This will be easy in the last constraint, because it just makes the right smaller,
but will be harder for the middle constraint as the left gets smaller. We can conclude that
the monopolist will increase Fa until the middle constraint is also an equality:
βψ(xb ) − Fb = 0,
αψ(xa ) − Fa = αψ(xb ) − Fb ,
βψ(xb ) − Fb ≥ βψ(xa ) − Fa .
We have now done with manipulation of the constraints and we will consider the monop-
olist’s optimisation.
4.2.4. Writing Down the Monopolist’s Optimisation. Now we will consider how the monop-
olist will choose the menu ((xa , Fa ); (xb , Fb )) to maximise its profits. This is the problem
max φ(Fa − cxa ) + (1 − φ)(Fb − cxb ), subject to
((xa ,Fa );(xb ,Fb ))
βψ(xb ) − Fb = 0
αψ(xa ) − Fa = αψ(xb ) − Fb
βψ(xb ) − Fb ≥ βψ(xa ) − Fa
We have already simplified this problem by implementing the reduction in the constraints
we talked about in the previous section. Now let us use the 2 equality constraints to
6. ADVERSE SELECTION 15
We have got rid of all the constraints and we now have one very simple condition.
4.2.5. Solving the Monopolist’s Optimisation. We now have a simple inequality constrained
optimisation that tells us what the monopolist wants to do. Let us write down the La-
grangean for this problem
L(xa , xb ) = φαψ(xa ) + (β − φα)ψ(xb ) − φcxa − (1 − φ)cxb + λ(xa − xb ).
Finding the first order conditions.
∂L
= φαψ 0 (xa ) − φc + λ = 0,
∂xa
∂L
= (β − φα)ψ 0 (xb ) − (1 − φ)c − λ = 0.
∂xb
Now we need to worry about whether the constraint binds or not. Suppose the constraint
was binding (λ > 0) and xa = xb = x then if these two equations are added together we
get βψ 0 (xa ) = c. Then substitute this into the first equation to get
φ(α − β)ψ 0 (xa ) + λ = 0,
which is impossible, because everything on the left is positive. This contradiction tells us
the constraint cannot be binding and λ = 0.
16 MARTIN CRIPPS
Substituting the fact that the constraint is not binding into the first order conditions we
get:
β(1 − φ)
αψ 0 (xa ) = c, βψ 0 (xb ) = c > c.
β − φα
So we learn that the high value types get sold their optimal quantity and the low value
types get sold lessC.#Two'Part#Tariffs#and#Adverse#Selection7
than their optimal quantity.
#7
At"any"given"price"the"
αψ`(.)" type"α’s"will"buy"more"
and"get"more"value."
βψ`(.)"
c(1 φ)/[1 φ(α/β)]
c"
x"
xb " x a"
What fees do they pay? We substituted these out of the problem with the substitutions
Fb = βψ(xb ) and Fa = αψ(xa ) + (β − α)ψ(xb ). The first of these, Fb = βψ(xb ), says the
C.#Two'Part#Tariffs#and#Adverse#Selection7
low-value types get not extra value. The fee they pays takes all their utility.
#7
αψ`(.)"
βψ`(.)"
c(1 φ)/[1 φ(α/β)]
c"
x"
xb " x a"
But notice the high-value types don’t pay all their utility to the monopolist. They pay
less than their total utility: Fa = αψ(xa ) − (α − β)ψ(xb ). This is because they could
pretend to be low-value types. If they make this pretence they get strictly positive utility
αψ(xb ) − Fb = (α − β)ψ(xb ) > 0. So this is the amount of utility they get to keep.
6. ADVERSE SELECTION 17
To conclude: The right way to solve the adverse selection problem for the monopolist is
to efficiently serve the high-value customers and treat the low-value types suboptimally.
7. MORAL HAZARD AND INCENTIVES
MARTIN CRIPPS
1. Introduction
Economic relationships often have the form of a Principal who contracts with an Agent
to take certain actions that the principal cannot observe directly. In each of these cases the
principal wants to control what the agent does, because this affects the principal and the
agent’s payoffs, but has limited ability to directly control the agent’s actions. These are
all examples of what economists call Moral Hazard. We think of moral hazard as arising
in situations where individuals have hidden or secret actions they can take.
For there to be an economic problem here it is necessary that three conditions hold: (1)
The Principal and Agent have different payoffs/objectives. (2) The Agent has actions it
can take. (3) The actions the Agent takes are not directly monitored by the Principal. If
any one of the three conditions fail, then there is not problem. If (1) is not true then the
Principal can rely on the Agent to take actions they both like. If (2) is not true then there
is nothing the Agent can do that will harm the Principal. And if (3) is not true then the
Principal can choose to only reward the Agent if they take the Principal’s most preferred
action.
If an incentive problem/moral hazard problem exists there will be a loss in efficiency.
The Principal will try to devise a reward/incentive scheme that will encourage the Agent
to take actions the Principal likes. These incentives impose an economic cost and harm
efficiency. We are all familiar with examples of bad incentive schemes. Here are two
examples:
• 1960’s: dentists in the UK health service had a government contract that paid
them for every cavity/filling they drilled independently of whether the tooth needed
drilling.
1
2 MARTIN CRIPPS
• Cost Plus Contracts: A Principal will reimburse a builder for allowable costs and a
provision for normal rates of profit (a cost plus contract) on a construction project.
However, even the very best incentive scheme will still impose economic costs and generate
inefficiency!
1.1. Adverse Selection or Moral Hazard? Under adverse selection there is hidden
information and under moral hazard there are hidden actions. Sometimes it is difficult
to tell the difference between moral hazard and adverse selection. Here is an example.
Suppose it was observed that Volvos go through stop signs more than other makes of car.
This has two potential explanations one is about moral hazard: “Volvo drivers believe
they?re safer and take less care when they drive as a result”. The other is about adverse
selection: “Bad drivers tend to buy Volvos.” Clearly both might be true but there are very
different implications for these two explanations of the same fact.
1.2.2. High-Powered Incentives. This is when a small change in observable outcome has a
big change in the Principal’s decision. You see these when it is particularly important to
avoid a bad outcome: Pass-Fail Grading Schemes, Sentences for Serious Crimes, Certifica-
tion of Nuclear Power Plants are all examples of high-powered incentives.
1.2.3. Relative Performance Evaluation. When good performance by the agent is partic-
ularly difficult to measure, incentives will compare the performance of one agent with
another agent to judge good performance. Tournaments, Benchmarking and Competition
are all examples of relative performance evaluation.
1.2.4. Target and Penalty Schemes. These are schemes that award a bonus when a certain
level of performance is achieved. The bonus has the effect of discouraging further effort
and there are mild penalties for under-perfomance—these schemes provide low-powered
incentives.
1.2.5. Efficiency Wages. Firms sometimes choose to pay their workers more than the mar-
ket wage. If they paid exactly what a worker could get elsewhere, then the worker does
not care if she is fired or not. (She can immediately obtain the same wage elsewhere.) The
more a worker is paid in his current job relative to the market wage, the more costly it is
for her to get fired (and she doesn’t want to shirk). Companies want to raise wages and
perpetuate unemployment as this gives workers incentives to try hard in their current job.
7. MORAL HAZARD AND INCENTIVES 3
1.3. Ownership: One Solution to the Incentive Problem. We’ve seen that incentive
problems arise if 3 conditions hold: 1. There is a divergence of interests. 2. There is a
need for the individuals to transact. 3. There are observation problems. One solution to
this problem is to sell the issue to the agent. Once the worker owns the company, they will
of course be willing to put in effort. Thus they should be willing to buy the company from
its owners at a price that correctly reflects its full value.
There are 2 problems with this argument. First the agent may not be able to borrow
sufficient money to acquire the company (we say the agent is credit-constrained). Second,
the company might be risky and the agent may care more about the risks of owning the
company than the current owners do.
1.4. Incentives and Risk. In a situation of moral hazard, the Principal observes an
outcome that is correlated with the action that she wants to encourage and rewards the
agent based on this outcome. Usually, this outcome has a random element in it. Sometimes
the agent puts in effort but is just unlucky and a bad outcome occurs. This randomness in
the outcomes makes the agent’s life risky because the rewards they get from the Principal
are random and usually individuals don?t like this risk.
Thus incentives introduce an inefficiency because they force Agents (workers, students
etc.) to bear risk. Although it is usually better for the Principal (firms, government etc.)
to bear the risk not the individual agents. It is actually essential for agents to bear some
risk, because if they experience no risk there are no incentives! So as economists we want
individuals to bear some risk (to give them incentives) but not too much (to impose too
great costs on them). Also, they might simply refuse to participate in schemes that are
too risky. (Of course, if the agent is risk-neutral and doesn’t care about risk this is not a
problem.)
1.5. Summary. In setting up incentives the general message is you must balance several
forces. They are (1) The increased benefit from better behaviour from agents. (2) The costs
of risk borne by agents (risk aversion). (3) How precisely you can measure performance.
(4) How much effort will increase in response to incentives anyway!
Incentive Intensity Principle: Incentives should be most intense when agents are able to
take actions to respond to them.
2.1. The Model. A company has £x in collateral, but needs a sum of money £1 > £x to
finance a risky investment project. The project has returns that are risky and also depends
on the effort put in by the company. If the company puts in no effort the project will pay
out
£3 probability = 1/3; £0 probability = 2/3.
If the company puts in high effort the project will pay
High effort costs the company £0.8 in lost opportunities elsewhere but low effort is free.
2.2. Optimal Decision When Investment is Self Financed. If the investment is made
from the company’s own assets, the company’s profits from its two effort levels are:
It is efficient in this case for the project to go ahead and for the company to put in high
effort.
2.3. Optimal Decision When Investment is Bank Financed. Now suppose the com-
pany borrows all the cost of the investment from the bank and must pay the £1 back from
any money it makes on the project. Then the above calculation becomes:
In this case the company is better off putting in low effort (this save it the costs of effort).
Notice that when the project doesn’t succeed the loan is not repaid, so this doesn’t affect
the company’s profits. (This is the company declaring bankruptcy.) Essentially, what
happens when the investment is borrowed is that the company gets to gamble with the
bank’s money and would rather not put in any effort of its own. The borrower here bears
no real costs of risk.
2.4. Optimal Decision When Company Puts in Some Collateral. Suppose the
company puts in £x in collateral and borrows 1 − x from the bank. Then the profits from
high and low effort are
Suppose he bank will only loan the funds if it knows the company will put in high effort,
because this maximises the probability of it getting its loan repaid. What value of collateral
7. MORAL HAZARD AND INCENTIVES 5
3.1. The Model. The worker can choose either to put in low effort e = 0 or high effort
e = 1. The amount of effort the agent puts in affects how much output q gets produced.
Here is the relationship between effort and output:
e = 0, ⇒ q = 0;
(
q = 1 prob = π
e = 1, ⇒
q = 0 prob = 1 − π
Thus when output is produced (q = 1) the manager knows that the worker put in high
effort, but if no output is produced (q = 0) the manager does not know whether the worker
put in high effort but was unlucky or just put in low effort (shirked).
We will assume that e = 1 is costly for the worker and costs them c > 0, but e = 0 is
costless for the worker. We will also assume the worker can go elsewhere and get utility
U > 0, so it doesn’t have to work for the firm and cannot be forced to have arbitrarily bad
utility.
The manager is able to provide an incentive scheme for the worker. In this case she
can see what output the worker produces but not see the effort they put in. So we allow
the manager to choose an employment contract that pays two different wages (w0 , w1 ),
where w0 is the wage the worker gets paid if they produce nothing and w1 is the wage the
worker gets if they produce some output. Thus the worker has a payoff that depends on
its expected wage net of its effort costs. There are three different actions he can take and
6 MARTIN CRIPPS
Now let us think about the profits of the firm (what the manager gets). These will
depend on how much output gets produced and how much wages are paid. We will assume
that output can be sold at the price of one, so profit equals 1q − w = q − w. Again the
manager’s expected profit will depend on the worker’s action:
This completes the description of the model, now we proceed to solve it. . .
3.2. The Worker’s Optimal Behaviour. As usual we work backwards. We first under-
stand what the worker will do for various different wage contracts (w0 , w1 ) and then use
this to understand what the right choice of contract is for the manager. We already have
figured out what the worker will get from each of his 3 actions. These are the expressions
(1),(2),(3) above. For the worker to be willing to come and work for the firm he must do
better at the firm than taking his outside option, that is, one of (1),(2) must be bigger
than (3).
So if one of: πw1 + (1 − π)w0 − c ≥ U or w0 ≥ U then the worker is better off working for
the firm than not. This is called the individual rationality constraint, or the participation
constraint, and is analogous to the IR constraint in adverse selection models.
If the worker works for the firm and prefers to provide high effort (e = 1) then his utility
from high effort (1) must be greater than his utility from low effort (2). This gives us a
second constraint
which is called the incentive compatibility constraint. This says the wage contract gives
the worker the incentive to put in high effort.
If the manager of the firm wants the worker to work and to provide high effort then two
constraints must hold:
πw1 + (1 − π)w0 − c ≥ U IR
πw1 + (1 − π)w0 − c ≥ w0 IC
(U+c)/π
Individual#Ra.onal#
Contracts#
w1#=###π:1#(U+c)#:##π:1#(1:π)#w0##
c/π w1#=#π:1(c+w0)#:#π:1(1:π)#w0#
(U+c)/(1:π)
w0#
But if the manager wants the worker to provide low effort then the constraints are
w0 ≥ U IR
w0 ≥ πw1 + (1 − π)w0 − c IC
3.3. The Manager’s Problem. The manager wants to maximise the profits made. We
have written profits down in the expressions (4) and (5). Suppose the manager wanted the
worker to provide high effort then she would choose a contract (w0 , w1 ) to solve the profit
maximisation problem
We have a picture of the constraints on this optimisation above. What about the manager’s
objectives. Well they want to maximise the profits which is equivalent to minimising the
expected wage they have to pay. Thus there profits increase as the expected wage decreases.
8 MARTIN CRIPPS
C.#A#Labour#Contracting#Model#4
w1#
Manager’s#indifference#curves#go#up#
#4 as#expected#wage#falls#
Constant =πw1#+(1-π)w0#
Higher#profit#and#
lower#costs#in#this#
direcAon#
w0#
If we now combine the objectives of the manager with the constraint set above we can see
that the manager would maximise her profit by choosing a wage contract on the yellow
segment of the figure below.C.#A#Labour#Contracting#Model#4
w1#
#4 There#are#a#range#of#
solu8ons#to#this#
problem.#
(U+c)/π
c/π
(U+c)/(1+π)
w0#
This line segment has the endpoints (w0 , w1 ) = (0, (U +c)/π) and (w0 , w1 ) = (U, U +(c/π)).
Any convex combination of these two points is an optimal contract. The expected wage
that is paid by the manager at the optimal contract is
U +c c
(1 − π)0 + π = (1 − π)U + π(U + = U + c.
π π
The manager thus gets a maximised profit equal to π − U − c.
If π > U + c the manager could make positive profit from this type of contract. Notice
that this is fully efficient. If the expected output π is greater than the social costs of high
effort c + U (where U is the opportunity cost of the worker coming to the firm), then it
is economically efficient for the worker to provide high effort. Thus there appear to be no
efficiency costs of incentives in this model.
7. MORAL HAZARD AND INCENTIVES 9
The expected wage above barely satisfies the workers IR constraint. (The worker is just
indifferent between working for the firm or giving up and working elsewhere.) The situation
is different for the IC constraint. At the contract (w0 , w1 ) = (U, U +(c/π)) the worker is just
indifferent between supplying high effort and shirking. However, at all other contracts the
worker strictly prefers higher effort. For example, at the contract (w0 , w1 ) = (0, (U + c)/π)
the wage for no output is so low that the worker strictly prefers to provide high effort
rather than low effort. Notice that this wage contract is very risky. The worker only gets
paid if they give high effort. As the worker only cares about his expected wage, this is
not a problem here. But if the worker disliked risk, then they would prefer the contract
(w0 , w1 ) = (U, U + (c/π)) to the contract (w0 , w1 ) = (0, (U + c)/π).
Suppose the manager decided to have low effort from the worker, then they would get
zero output (from the low effort) and would want to minimise their wage bill. To do this
they could always pay zero (and encourage the worker to go elsewhere).
4.1. The Model. The model is very similar to the one in the previous section again there
will be two effort levels and two output levels. Now the effort level will be called e− (low
effort) and e+ (high effort) and the output levels will be called y − (low output) and y +
(high output). When the worker puts in low effort, e− , the probability of high output
is π(e− ). And when the worker puts in high effort, e+ , the probability of high output is
π(e+ ). Of course we will assume π(e+ ) > π(e− ). In summary we have:
(
− y = y − prob = 1 − π(e− )
e=e ⇒ ;
y = y + prob = π(e− )
(
y = y − prob = 1 − π(e+ )
e = e+ ⇒ .
y = y + prob = π(e+ )
In contrast to the previous section, it is possible for the worker to produce high output
even when they put in low effort. Thus, the manager is never certain what the worker
did. The difference π(e+ ) − π(e− ) measures how effectively the manager can monitor the
worker. If π(e+ ) − π(e− ) is zero, it is impossible for the manager to ever know anything
about the worker’s behaviour or to provide incentives. If π(e+ ) − π(e− ) = 1 the manager
perfectly knows what the worker has done.
As before we will write a wage contract as a pair (w− , w+ ), where w− is the wage the
worker gets after low output is realised and w+ is the worker’s wage after high output is
achieved. The worker has a utility function u(w) and a cost of effort function c(e), so we
10 MARTIN CRIPPS
can write down the worker’s payoffs to the three different actions they can take.
4.2. The Constraints. We will focus on contracts where the worker provides high effort.
In this case the individual rationality constraint only requires that the worker prefers high
effort to quitting
The IR constraint cannot be manipulated in a useful way. That is not true of the IC
constraint. If the worker is to provide high effort he must also prefer high effort to low
effort that is
π(e+ )u(w+ ) + (1 − π(e+ ))u(w− ) − c(e+ ) ≥ π(e− )u(w+ ) + (1 − π(e− ))u(w− ) − c(e− ). IC
π(e+ )u(w+ ) + (1 − π(e+ ))u(w− ) − c(e+ ) ≥ π(e− )u(w+ ) + (1 − π(e− ))u(w− ) − c(e− )
π(e+ )[u(w+ ) − u(w− )] − c(e+ ) ≥ π(e− )[u(w+ ) − u(w− )] − c(e− )
π(e+ )[u(w+ ) − u(w− )] − c(e+ ) ≥ π(e− )[u(w+ ) − u(w− )] − c(e− )
[π(e+ ) − π(e− )][u(w+ ) − u(w− )] ≥ c(e+ ) − c(e− )
This says the increase in the probability of a high wage times in the increase in the utility
of the high wage must be greater than the increase in the cost of a high wage.
4.3. The Firm’s Objectives. If the worker puts in high effort, there is probability π(e+ )
of a high output and the firm then has to pay a high wage and there is probability 1−π(e+ )
of low output and the firm then has to pay a low wage. Thus the firm’s expected profits,
when the worker puts in high effort is
So, if we assume that high effort is worth paying for, then the firm?s profit maximisation
problem boils down to maximising the above expression (by choosing wages) subject to the
two constraints in the previous section. That is
4.4. Solving the Firm’s Problem. We now really need to use our Lagrangean technique
to solve the firm’s problem. It has two constraints and it wants to choose a wage contract to
maximise its expected profits. We are going to have to write down a rally big Lagrangean:
L(w+ , w− ) = π(e+ )(y + − w+ ) + (1 − π(e+ ))(y − − w− )
+ λ π(e+ )u(w+ ) + (1 − π(e+ ))u(w− ) − c(e+ ) − U
Then differentiate it with respect to w+ and w− . You can see that each of these two
variables appear in many places here. There are 3 different places where w− appears this
gives
∂L
= (1 − π(e+ ))(−1) + λ(1 − π(e+ ))u0 (w− ) + µ[π(e+ ) − π(e− )][−u0 (w− )] = 0
∂w−
If we divide this equation though by u0 (w− )(1 − π(e+ )) it becomes
−1 π(e+ ) − π(e− )
+ λ − µ =0
u0 (w− ) 1 − π(e+ )
Or
π(e+ ) − π(e− ) 1
λ−µ +
= 0 −
1 − π(e ) u (w )
Now let us differentiate the Lagrangean with respect to w+ . This gives
∂L
= π(e+ )(−1) + λπ(e+ )u0 (w+ ) + µ[π(e+ ) − π(e− )]u0 (w+ ) = 0
∂w−
Now we divide this through by π(e+ )u0 (w+ ) to get
−1 π(e+ ) − π(e− )
+ λ + µ =0
u0 (w+ ) π(e+ )
Or
π(e+ ) − π(e− ) 1
λ+µ +
= 0 +
π(e ) u (w )
We will now start to do some thinking about these conditions and what the tell us. Let
us start by writing them in a slightly different way:
1
u0 (w− ) = π(e+ )−π(e− )
λ − µ 1−π(e+ )
1
u0 (w+ ) = + −) .
λ + µ π(e π(e
)−π(e
+)
Suppose that λ > 0 and µ > 0 (we will address this issue in a little while). Then 1/(λ−. . . )
is bigger than 1/(λ+. . . ), so we know that the the marginal utility at the wage w− is greater
than the marginal utility at the wage w+ . This means that w+ > w− . The extent to which
this is true depends on the shape of the utility function and how quickly marginal utility
varies with the wages.
12 MARTIN CRIPPS
The gap between the two wages determines the strength of the incentives in this model.
What makes w+ − w− big? Making µ or π(e+ ) − π(e− ) big will strengthen incentives,
because it is this that makes the denominators in these two fractions far apart.
π(e+ ) − π(e− ) π(e+ ) − π(e− ) [π(e+ ) − π(e− )][1 − 2π(e+ )]
λ+µ − λ − µ = µ
π(e+ ) 1 − π(e+ ) (1 − π(e+ ))π(e+ )
But this makes perfect sense when µ is big the constraint on getting high effort is very
important so it makes sense for strong incentives to be provided. Also when π(e+ ) − π(e− )
is big the firm gets very accurate information from its output signals so strong incentives
are powerful.
Technical Issue: Now we might start to worry about the Lagrange multipliers being zero.
Suppose, λ = 0 then the first order condition for w− becomes
π(e+ ) − π(e− ) 1
−µ +
= 0 −
1 − π(e ) u (w )
but this is impossible because it says marginal utility is negative. So we can conclude the
λ > 0. Now suppose µ = 0. If this is substituted into the first order conditions we get
Then we have to have u0 (w1 1 + = w − , because marginal utility
− ) = u0 (w + ) = λ. This says w
increases in w. But it is impossible to satisfy the IC constraint if the same wage is paid
for both output levels.
ECON2001 7. DESIGNING ECONOMIC SYSTEMS
MARTIN CRIPPS
V=40 V=65
C=35 40-35=5 65-35=30
C=60 0 65-60=5
When C = 60 and V = 40 it is inefficient for the buyer and seller to trade, however, in all
other cases trade is efficient. The total expected increase in value that efficient trade can
1
2 MARTIN CRIPPS
generate is
1 1 1 1
(1) 10 = 0 + 5 + 5 + 30.
4 4 4 4
Thus if we could organise trade in all the cases where trade should occur, then welfare
could be increased by 10, however, we will no establish (in this example) the following
result:
Result 1 (Myerson Satterthwaite). There is no way of organising trade in this example
under which the project gets built if and only if it is efficient to do so.
1.1. Split the Difference Doesn’t Work. Suppose you tried to organise trade by getting
the buyer and seller to tell you their values and then choosing a price that would ensure
they shared the gains from trade equally. Then the prices you would choose are
V=40 V=65
C=35 P=37.5 P=50
C=60 P=62.5
(The missing entry in this table would be where no trade would occur.) Would the seller
want to tell the truth in this scheme? If the seller was low cost and truthfully announced her
type, she could expect to trade at the price P = 37.5 half the time and at the price P = 50
half the time, making an expected profit of (37.5 − 35)0.5 + (50 − 35)0.5 = 8.75. But if she
lied and said her costs were high she would trade half the time at the price 62.5 and not
trade the rest of the time. This would give her the expected profit 0.5(62.5 − 35) = 13.75.
Thus, she would be better off lying and not telling the truth and the split the difference
scheme does not work.
As an exercise check whether the high cost seller wants to tell the truth and whether
the buyer’s two types want to tell the truth here.
1.2. Myerson Satterthwaite in this Example. Now we will give an argument to show
that it is impossible to get the buyer and the seller to truthfully reveal their type in this
example. Thus it is impossible to organise trade efficiently. The reason this argument works
is that it is necessary to give the buyer and seller incentives to tell the truth, because you the
organiser needs to know when trade should occur and when it should not. These incentives
are costly. The total amount of value that can be created from trade is insufficient to
provide these incentives.
First notice that the total amount of surplus created from trade is 10, by the calculation
(1) above. Now suppose we begin by thinking about the high-cost seller. To get the high-
cost seller to trade with the high-value buyer we have to promise them a price of at least
60 half the time. But this means that the low-cost seller, by pretending to be the high-cost
seller, can get at least 25 = 60 − 35 half the time. So the very least the low-cost seller can
get from cheating on truthtelling is 0.5(60 − 35) = 12.5. But we want the low-cost seller
to tell the truth. So we must promise the low-cost seller at least 12.5 if they are to tell the
truth. Half of the time the seller is high cost (and we must give them some value) half of
ECON2001 7. DESIGNING ECONOMIC SYSTEMS 3
the time the seller is low cost and we must give them at least 12.5. So we must expect to
give at least
1 1
0 + 12.5 = 6.25
2 2
of the surplus from trade to the seller.
Now suppose we begin by thinking about the low-value buyer. To get the low-value
buyer to trade with the low-cost seller we have to promise them a price of at most 40 half
the time. But this means that the high-value buyer, by pretending to be the low-value
buyer, can get at least 25 = 65 − 40 half the time. So the least the high-value buyer can
get from cheating on truthtelling is 0.5(65 − 40) = 12.5. But we want the high-value buyer
to tell the truth. So we must promise the high-value buyer at least 12.5 if they are to tell
the truth. Half of the time the buyer is low value (and we must give them some surplus)
half of the time the buyer is high value and we must give them at least 12.5. So we must
expect to give at least
1 1
0 + 12.5 = 6.25
2 2
of the surplus from trade to the buyer.
Combining these two calculations we have to give at least 6.25 to the seller to get
truthtelling and at least 6.25 to the buyer to get truthtelling and there is only 10 in surplus
to be shared. It is clearly impossible to get truthtelling because it requires resources of 12.5
and there is only 10 available. This result has profound implications for many economic
arguments. The idea that individual bargaining or negotiation leads to efficiency is not
true when the bargainers have private (payoff relevant) information. Arguments such as the
Coasian solution to the externality problem, or many Chicago-style analyses are vulnerable
to this issue.
If we look at the ends of this process, ignoring the intermediate step, we get a relationship
between the players’ types and the decision that gets taken
(θ1 , . . . , θN ) 7→ γ(s∗1 (θ1 ), . . . , s∗N (θN ))
(θ1 , . . . , θN ) 7→ Γ(θ1 , . . . , θN )
This tells us that if the players are the type (θ1 , . . . , θN ) it is possible to take the deci-
sion Γ(θ1 , . . . , θN ) ≡ γ(s∗1 (θ1 ), . . . , s∗N (θN )). This is called Implementation. That is, there
exists a mechanism that allows you to take the decision Γ(θ1 , . . . , θN ) when the types are
θ1 , . . . , θ N .
For example, one decision rule we would like to implement is the rule that obliges the
buyer and seller in to trade whenever it is efficient. We know this is impossible, by the
Myerson & Satterthwaite result. Another decision rule is to oblige the buyer and seller to
trade only if the buyer is high value and the seller is low cost. We know this is possible—just
choose an appropriate fixed price.
The main result we have on implementation is that in many situations we do not need
to give the agents complicated message spaces to get optimal behaviour. The Revelation
Principle states that if the strategies s∗i (θ) are an equilibrium in dominant strategies. It
is enough to allow the agents to just tell you their type Si = Θi . No more complicated
messages are necessary. The idea behind this is to take any arbitrary mechanism with an
equilibrium (s∗1 (θ1 ), . . . , s∗N (θN )) and add on to it an initial stage where players are asked
to report their type and if they report θi the initial stage then sends the message s∗i (θi ) to
the mechanism. This will not change the players’ incentives if (s∗1 (θ1 ), . . . , s∗N (θN )) was an
equilibrium of the original mechanism.
If U > C. Then the project will be built whatever individual j reports, they don’t care
what they report, so it is optimal for individual j to report truthfully.
Now suppose that U + vj > C > U . In this case reporting truthfully ṽj = vj , will result
in the project being built and individual j is pivotal. They will then have to pay C − U ,
thus reporting truthfully gets them the utility vj − (C − U ) > 0 (as vj + U > C). But if
they understate their value the project will not be built and they get value zero. Truthful
reporting is better than under reporting.
Finally C > U + vj . In this case reporting truthfully results in the project not being built
and the individual getting zero utility. Exaggerating the value and reporting ṽj > C − U
will get the project built, but because the individual is pivotal they must pay C − U which
is greater than their true value.
4. Auctions
Auctions are a very popular way of selling an object when the seller does not know
the demand function. In the classical model of monopoly a monopolist knows its demand
function and should just set an optimal price (this is called the Posted-Price Model).
When a monopolist doesn?t know demand it may make a big mistake by behaving in this
way (and not sell or sell too cheaply). Having an auction allows you to learn demand
(by observing the bids) and set a good price at the same time. Thus the monopolist is
finding a mechanism to reveal its demand and to determine its price. The idea here is that
competition amongst buyers gets them to tell you the true demand.
ECON2001 7. DESIGNING ECONOMIC SYSTEMS 7
At an equilibrium the bid b∗ (v) is optimal for type v so pretending to be v 0 and bidding
according to this this has to be worse than bidding b∗ (v). That is
F (v)N −1 [v − b∗ (v)] ≥ F (v 0 )N −1 [v − b∗ (v 0 )] for all v 0 .
Another way of saying the same thing is to say the right hand side of the inequality above
is maximized at v 0 = v.
v0 = v Maximises F (v 0 )N −1 [v − b∗ (v 0 )].
If we differentiate this and set it equal to zero we will find the maximum
d 0 N −1 ∗ 0 0 0 N −2 ∗ 0 0 N −1 db∗ 0
F (v ) [v − b (v )] = (N − 1)f (v )F (v ) [v − b (v )] + F (v ) [− (v )]
dv 0 dv
Now setting the derivative equal to zero and v 0 = v we get the differential equation
db∗
0 = (N − 1)f (v)F (v)N −2 [v − b∗ (v)] − F (v)N −1
dv
A little re-arranging gives
db∗
F (v)N −1 + (N − 1)f (v)F (v)N −2 b∗ (v) = (N − 1)f (v)F (v)N −2 v.
dv
The left hand side of this is just the derivative of a product
d
F (v)N −1 b∗ (v) = (N − 1)f (v)F (v)N −2 v
dv
So integrating both sides allows us to solve for the optimal bidding function
Z
N −1 ∗
F (v) b (v) = (N − 1)f (v)F (v)N −2 vdv
or
(N − 1)f (v)F (v)N −2 vdv
R
∗
b (v) = .
F (v)N −1
This expression tells us that the optimal bid in a first price auction is the bidders best
guess about the second highest bid.