Professional Documents
Culture Documents
1 Introduction To Structure of Dynamic Oligopoly Models: 1t 2t 1t 2t
1 Introduction To Structure of Dynamic Oligopoly Models: 1t 2t 1t 2t
t=0
t
(x
1t
, x
2t
, q
1t
, q
2t
)
where ( ) denotes single-period prots.
Because the two rms are duopolists, and they must make these choices recog-
nizing that their choices can aect their rivals choices. We want to consider
a dynamic equilibrium of such a model, when (roughly speaking) each rms
sequence of qs is a best-response to its rivals sequence.
A rms strategy in period t, q
it
, can potentially depend on the whole history
of the game (H
t1
{x
1t
, x
2t
, q
1t
, q
2t
}
t
=0,... ,t1
), and well as on the time period
t itself. This becomes quickly intractable, so we usually make some simplifying
regularity conditions:
Firms employ stationary strategies: so that strategies are not explicitly a
function of time t (i.e. they depend on time only indirectly, through the
history H
t1
). Given stationarity, we will drop the t subscript, and use
primes
to denote next-period values.
A dimension-reducing assumption is usually made: for example, we might
assume that q
it
depends only on x
1t
, x
2t
, which are the payo-relevant
Lecture notes: second set 2
state variables which directly aect rm is prots in period i. This is usu-
ally called a Markov assumption. With this assumption q
it
= q
i
(x
1t
, x
2t
),
for all t.
Furthermore, we usually make a symmetry assumption, that each rm
employs an identical strategy assumption. This implies that q
1
(x
1t
, x
2t
) =
q
2
(x
2t
, x
1t
).
To characterize the equilibrium further, assume we have an equilibrium strategy
function q
(x
1
, x
2
) = argmax
q
{(x
1
, x
2
, q, q
(x
2
, x
1
)) + V (x
1
= g (x
1
, q) , x
2
= g (x
2
, q
(x
2
, x
1
)))}
(1)
from rm 1s perspective, and similarly for rm 2. V (, ) is the value function,
dened recursively at all possible state vectors x
1
, x
2
via the Bellman equation:
V (x
1
, x
2
) = max
q
{(x
1
, x
2
, q, q
(x
2
, x
1
)) + V (x
1
= g (x
1
, q) , x
2
= g (x
2
, q
(x
2
, x
1
)))} .
(2)
I have described the simplest case; given this structure, it is clear that the
following extensions are straightforward:
Cross-eects: x
i
= g (x
i
, x
i
, q
i
, q
i
)
Stochastic evolution: x
i
|x
i
, q
i
is a random variable. In this case, replace
last term of Bellman eq. by E [V (x
1
, x
2
) |x
1
, x
2
, q, q
2
= q
(x
2
, x
1
)].
This expectation denotes player 1s equilibrium beliefs about the evolution
of x
1
and x
2
(equilibrium in the sense that he assumes that player 2 plays
the equilibrium strategy q
(x
2
, x
1
)).
> 2 rms
Firms employ asymmetric strategies, so that q
1
(x
1
, x
2
) = q
2
(x
2
, x
1
)
. . .
Lecture notes: second set 3
Computing the equilibrium strategy q
( ),
via the assumption that the rival rm is always using the optimal strategy.
So value iteration procedure is more complicated:
1. Start with initial guess V
0
(x
1
, x
2
)
2. If qs are continuous controls, we must solve for q
0
1
q
0
(x
1
, x
2
) and
q
0
2
q
0
(x
2
, x
1
) to satisfy the system of rst-order conditions (here
subscripts denotes partial derivatives)
0 =
3
_
x
1
, x
2
, q
0
1
, q
0
2
_
+ V
0
1
_
g
_
x
1
, q
0
1
_
, g
_
x
2
, q
0
2
__
g
2
_
x
1
, q
0
1
_
0 =
3
_
x
2
, x
1
, q
0
2
, q
0
1
_
+ V
0
1
_
g
_
x
2
, q
0
2
_
, g
_
x
1
, q
0
1
__
g
2
_
x
2
, q
0
2
_
.
(3)
For the discrete control case:
q
0
= argmax
qQ
_
_
x
1
, x
2
, q, q
0
2
_
+ V
0
_
g (x
1
, q) , g
_
x
2
, q
0
2
___
q
0
2
= argmax
qQ
_
_
x
2
, x
1
, q, q
0
1
_
+ V
0
_
g (x
2
, q) , g
_
x
1
, q
0
1
___
.
(4)
3. Update the next iteration of the value function:
V
1
(x
1
, x
2
) =
_
_
x
1
, x
2
, q
0
1
, q
0
2
_
+ V
0
_
g
_
x
1
, q
0
1
_
, g
_
x
2
, q
0
2
___
. (5)
Note: this and the previous step must be done at all points (x
1
, x
2
) in
the discretized grid. As usual, use interpolation or approximation to
obtain V
1
( ) at points not on the grid.
4. Stop when sup
x
1
,x
2
||V
i+1
(x
1
, x
2
) V
i
(x
1
, x
2
)|| .
In principle, one could estimate a dynamic games model, given time series
of {x
t
, q
t
} for both rms, by using a nested xed-point estimation algo-
rithm. In the outer loop, loop over dierent values of the parameters ,
and then in the inner loop, you compute the equilibrium of the dynamic
game (in the way described above) for each value of .
However, there is an inherent Curse of dimensionality with dynamic
games, because the dimensionality of the state vector (x
1
, x
2
) is equal to
Lecture notes: second set 4
the number of rms. (For instance, if you want to discretize 1000 pts in one
dimension, you have to discretize at 1,000,000 pts to maintain the same
neness in two dimensions!)
Some papers provide computational methods to circumvent this problem
(Keane and Wolpin (1994), Pakes and McGuire (2001), Imai, Jain, and
Ching (2005)). Generally, these papers advocate only computing the value
function at a (small) subset of the state points each iteration, and then
approximating the value function at the rest of the state points using values
calculated during previous iterations.
1.1 Dynamic games with incomplete information
Clearly, it is possible to extend the Hotz-Miller insights to facilitate estimation of
dynamic oligopoly models, in the case where q is a discrete control. Advantage,
as before, is that you can avoid numerically solving for the value function.
Data directly tell you: the choice probabilities (distribution of q
1
, q
2
|x
1
, x
2
);
state transitions: (joint distribution of x
1
x
2
|x
1
, x
2
, q
2
, q
2
).
Papers which develop these ideas include: Bajari, Benkard, and Levin (2008),
Pesendorfer and Schmidt-Dengler (2003), Aguirregabiria and Mira (2007), Pakes,
Ostrovsky, and Berry (2007). Applications include: Collard-Wexler (2006),
Ryan (2006), and Dunne, Klimer, Roberts, and Xu (2006). See survey in Pakes
(2008), section 3.
Here I outline the general setting, which can be considered a multi-agent version
of Rust setup. (Nitty-gritty details are omitted: you should be able to gure it
out!)
First-order Markov
Let
X and q denote the N-vector of rms state variables, and actions.
Symmetric model: rms are identical, up to the current values of their state
variables, and idiosyncratic utility shocks
Lecture notes: second set 5
Utilities: rm is current utility, given her action d
i
and her utility shocks
i,d
i
is:
U(
X, d
i
,
d
i
; ) +
i,d
i
.
All utility shocks are private information of rm i. Also, they are i.i.d. across
rms, and across actions (exactly like the logit errors in Rusts framework).
Because of these assumptions: each rms period t shocks aect only his period
t actions; they do not aect opponents actions, nor do they aect actions in
other periods t
= t.
In single-agent case, this distinction is nonsensical. But in multi-agent context,
it is crucial. Consider several alternatives:
All the shocks are publicly observed each period, but not observed by
econometrician: essentially, we are in the unobserved state variable
world of the last set of lecture notes, except that in multi-agent context,
each rms actions each period depend on all the shocks of all the rms in
that period. Likelihood function gets very complicated.
Shocks are private information, but serially correlated over time: now rms
can learn about their opponents shocks through their actions, and their
beliefs may evolve in complicated fashion. Also, rms may strategically
try to inuence their rivals beliefs about their shocks (signalling). Com-
plicated even from a theoretical point of view.
Shocks are unobserved by both rms and econometrician. In this case,
they are non-structural errors and are useless from the point of view of
generating more randomness in the model relative to the data.
In symmetric equilibrium, rms optimal policy function takes the form d
i
=
d
(
X,
i
). Corresponding choice probability: P(d
i
|
X) (randomness in
i
). These
can be estimated from data.
State variables evolve via Markovian law-of-motion
X
X,
d. This can also be
estimated from the data.
Lecture notes: second set 6
Hence, choice-specic value functions for rm i (identical for all rms) can be
forward-simulated:
V (
X, d
i
= 1; ) =E
vecd
i
|
X
u(
X, d
i
= 1,
d
i
; ) + E
X,
d=(d
i
,
d
i
)
E
|d
i
,
X
_
u(
X
,
d
; ) +
+E
|d
i
,
X
_
u(
X
,
d
; ) +
+
__
(6)
When
X is nite, the matrix representation of value function can also be used
here.
Predicted choice probabilities:
P(d
i
= 1|
X) =
exp(
V (
X, d
i
= 1; ))
exp(
V (
X, d
i
= 1; )) + exp(
V (
X, d
i
= 0; ))
.
Demand functions
Given prices p
k
t
, p
(k)
t+1
for all k = 1, . . . , K, each period there are K indierent
consumers
1
t
2
t
3
t
. . .
K
t
0, s.t.
Lecture notes: second set 9
The indierent consumers solve
k
t
p
k
t
+ p
(k)
t+1
=
k+1
k
t
p
k+1
t
+ p
(k+1)
t+1
, for k = 1, . . . , K 1
Consumer heterogeneity: is uniformly distributed.
Derive inverse demand function (subject to non-negativity constraints)
p
k
t
= (
k
k+1
)
_
1
1
M
k
r=1
x
r
t
_
+ p
(k)
t+1
+ p
k+1
t
p
(k+1)
t+1
Supply side
y
t
= [1, x
1
t
, . . . , x
K
t
]
iL
c
i
x
(i)
t
.
Period t prots for car i is
i
t
_
y
t
, y
t+1
, . . . , y
t+T
i
1
_
=
_
p
i
t
(y
t
, y
t+1
, . . . , y
t+T
i
1
) c
i
t
_
q
1
t
:
depends on past, current, future prodn of car i.
Important: dependence of current prots on future actions leads to a time-
consistency problem, which is absent from usual dynamic problems. Very
roughly, time-inconsistency implies that an agents optimal action in period t
diers depending on whether the agent is deciding in period t, or period t 1,
or period t 2, etc.
Think of durable goods monopoly: in period 1, his optimal period 2 price is the
monopoly price (because that would raise his prots in period 1). But when
period 2 comes, his optimal period 2 price is actually a lower price (since he
wants to sell to people who did not buy in period 1).
Lecture notes: second set 10
For individual rm: t, j N, i L
j
, period-t production x
(i)
t
maximizes
() max
x
(i)
t
,iL
j
=0
iL
j
i
t+
_
y
t+
, y
t++1
, . . . , y
t++T
i
1
_
. .
period t + prots
, s.t.
y
t+
= Ay
t+1
+Bd
t+
, for = 1, . . . , .
Note: obj fxn dierent in each period t: usual problem is
max
n
x
(i)
t
, iL
j
o
t=0
t=0
iL
j
i
t
_
y
t
, y
t+1
, . . . , y
t+T
i
1
_
.
FOC for x
(i)
t
contains derivative (say)
i
t1
x
(i)
t
: in choosing period-t prodn, rec-
ognize that it aects period-(t + 1) prots time-inconsistent.
iL
j
j
(y
t
, y
t+1
, . . . , y
t+T
i
1
) +V
j
(Ay
t
)
for all rms j, and the Markov decision rules
x
i
t
= g
i
(Ay
t1
), for all i L
j
.
Value function V
j
(Ay
t1
)=(*) (optimal prots from t onwards).
t
A
SAy
t
.
Bellman equation can be rewritten in matrix notation
y
t1
A
SAy
t1
=
max
x
i
t
, iJ
_
iJ
_
T
i
1
z=0
y
t+z
z
R
z
(i)
y
t
__
y
t
C
j
y
t
+y
_
A
S
j
A
y
t
Recursive substitution yields
y
t1
A
SAy
t1
=
max
x
i
t
, iJ
y
t
__
iJ
T
i
1
z=0
(A
)
z
[(I +BG)
]
z
z
R
v
z
(i)
_
C
j
+
_
A
S
j
A
_
y
t
max
x
(i)
t
, iJ
y
t
Q
j
y
t
.
j
_
Q
j
+Q
j
_
y
t
= B
j
_
Q
j
+Q
j
_
Ay
t1
+B
j
_
Q
j
+Q
j
_
Bd
t
= 0.
This system of FOCs corresponds to Eq. (3) in dynamic games handout.
Rearranging, we get:
d
t+h
= (WB)
1
(WA) y
t1
,
where W
j
B
j
_
Q
j
+Q
j
_
for each rm j and W [W
1
, . . . , W
N
]
.
Basis for estimating supply side of model.
Lecture notes: second set 12
Estimation
is not identied, set to constant.
To generate estimating equations, introduce shocks to rms marginal costs:
C(x
(i)
t
) = x
(i)
t
(c
i
+
it
) .
Assumptions:
t
[
1t
, . . . ,
Lt
]
t+1
= 0.
This implies that the realized demand function residual will also have mean zero,
conditional on formation known at time t:
0 = E
_
_
p
(i)
t
_
(i)
(i)+1
_
_
_
1
1
M
(i)
r=1
x
r
t
_
_
p
((i))
t+1
p
(i)+1
t
+p
((i)+1)
t+1
t
_
_
= E
_
_
_
1 L
1
_
_
p
(i)
t
p
(i)+1
t
_
(i)
(i)+1
_
_
_
1
1
M
(i)
r=1
x
r
t
_
_
t
_
_
These form the basis for the moment conditions which we use for estimation, which
we denote
T
().
Lecture notes: second set 13
GMM Estimation
Source of identication: co-movements time series of prices and production.
T
()
1
T
T
().
Nested GMM procedure: for each value of parameters , solve LQ dynamic
programming problem for coecients G() of optimal production rules. LQ
dynamic programming problem is very easy and quick to solve.
Results
Lecture notes: second set 14
References
Aguirregabiria, V., and P. Mira (2007): Sequential Estimation of Dynamic Discrete
Games, Econometrica.
Bajari, P., L. Benkard, and J. Levin (2008): Estimating Dynamic Models of Imper-
fect Competition, Econometrica.
Collard-Wexler, A. (2006): Demand Fluctuations and Plant Turnover in the Ready-
to-Mix Concrete Industry, manuscript, New York University.
Dunne, T., S. Klimer, M. Roberts, and D. Xu (2006): Entry and Exit in Geographic
Markets, manuscript, Penn State University.
Esteban, S., and M. Shum (2007): Durable Goods Oligopoly with Secondary Markets:
the Case of Automobiles, Forthcoming, RAND Journal of Economics.
Imai, S., N. Jain, and A. Ching (2005): Bayesian Estimation of Dynamic Discrete
Choice Models, Queens University, mimeo.
Keane, M., and K. Wolpin (1994): The Solution and Estimation of Discrete Choice
Dynamic Programming Models by Simulation and Interpolation: Monte Carlo Evidence,
Review of Economics and Statistics, 76, 648672.
Pakes, A. (2008): Theory and Empirical Work in Imperfectly Competitive Markets,
Fisher-Schultz Lecture at 2005 Econometric Society World Congress (London, England).
Pakes, A., and P. McGuire (2001): Stochastic Algorithms, Symmetric Markov Perfect
Equilibrium, and the Curse of Dimensionality, Econometrica, 69, 12611282.
Pakes, A., M. Ostrovsky, and S. Berry (2007): Simple Estimators for the Parameters
of Discrete Dynamic Games (with Entry/Exit Examples), RAND Journal of Economics.
Pesendorfer, M., and P. Schmidt-Dengler (2003): Identication and Estimation of
Dynamic Games, NBER Working Paper, #9726.
Ryan, S. (2006): The Costs of Environmental Regulation in a Concentrated Industry,
manuscript, MIT.