Professional Documents
Culture Documents
09 Limit Theorems
09 Limit Theorems
Convergence in Distribution
The previous de…nitions of convergence focus on the outcome sequences of a ran-
dom variable. Convergence in distribution refers to the probability distribution
of each element of the sequence directly.
De…nition. A sequence of random variables Y1 ; Y2 ; : : : is said to converge
in distribution to a random variable Y if
lim P Yn < c = P (Y < c)
n!1
The next result combines convergence in probability and distribution, and is used
repeatedly to derive the limit theory for our estimators.
Lemma (parts a and c are called Slutzky’s Theorem):
P D D
(a) Yn ! and Xn ! X ) Yn + Xn ! X +
P D P
(b) Yn ! 0 and Xn ! X ) Yn0 Xn ! 0
P D D
(c) An ! A and Xn ! X ) An Xn ! AX
P D D
(d) An ! A and Xn ! X ) Xn0 An 1 Xn ! X 0 A 1 X
Parts (c)-(d) require that X and A be conformable and (for part (d)) that A be
invertible.
D D
In particular, if X N (0; V ), then Yn +Xn ! N ( ; V ) and An Xn ! N (0; AV A0 ).
Also, if we set = 0, then part (a) implies
P D D
Yn ! 0 and Xn ! X ) Yn + Xn ! X:
P p
Let Zn = Yn + Xn , so Yn ! 0 (i.e. Zn Xn ! 0) implies that the asymptotic
p
distribution of Zn is the same as that of Xn . When Zn Xn ! 0 we sometimes
say that the two sequences are asymptotically equivalent (Zn Xn ) and write
a
Zn = Xn + oP
where oP is some suitable random variable (Yn here) that converges to zero in
probability.
We employ the lemma in a standard trick to derive the limit distribution of a
sequence of random variables. Suppose we wish to …nd the limit distribution of
D P
Yn0 Xn , where we know Xn ! X and Yn ! . Then we replace Yn with , and
determine the limit distribution of 0 Xn . Why does this work? Because Yn0 Xn
a
0
Xn , which is more easily seen in the equivalent statement
Yn0 Xn = 0
Xn + oP
0
where oP here is Yn Xn .
To test nonlinear hypotheses, given the asymptotic distribution of the estima-
tor, we need
Lemma (delta method): Let fBn gn 1 be a sequence of K-dimensional random
variables such that
P
Bn !
1 D
n 2 (Bn ) ! N (0; V ) :
@g ( )
G( ) = :
@ 0
Then
1 D
n 2 (g (Bn ) g ( )) ! N 0; G ( ) V G ( )0 :
1 1 1
Proof: A mean-value expansion yields n 2 g (Bn ) = n 2 g ( ) + n 2 G (B ) (Bn ),
P P
where Bn is the mean value between Bn and . Because Bn ! , Bn ! and,
P
by continuity, G (Bn ) ! G ( ). Thus
1 1 D
n 2 [g (Bn ) g ( )] = n 2 G (Bn ) (Bn ) ! G ( ) N (0; V ) :
The following strong LLN assumes that fYt g is i.i.d. but the variance does not
need to be …nite
Kolmogorov’s Second Strong LLN: Let fYt g be i.i.d. with EYt = . Then
a:s:
Yn ! .
Central limit theorems are a principal method of establishing convergence in dis-
tribution, as they govern the limit behavior of the p
di¤erence between Yn and E Yn
(which equals EYt if fYt g is i.i.d.) blown up by n. For i.i.d. sequences, the
only CLT we need is
Lindberg-Levy CLT: Let fYt g be i.i.d. with EYt = and V ar (Yt ) = . Then
1 X
n
p d
n Yn =p (Yt ) ! N (0; ) :
n t=1
p
We read the above CLT as: a sequence of random vectors n Yn converges
in distribution to a random vector whose distribution is N (0; ). To understand
how to relate the above CLT for random vectors, with the CLT for random scalars,
let be any vector of real numbers with the same dimension as Yt . Now f 0 Yt g is
a sequence of random variables with E ( 0 Yt ) = 0 and V ar ( 0 Yt ) = 0 . The
scalar version of Lindberg-Levy implies
p p d
n 0 Yn 0
= 0 n Yn ! N (0; 0 ) :
Yet this limit distribution is the distribution of 0 Z, where Z N (0; ), hence
the multivariate convergence in distribution theorem establishes the multivariate
Lindberg-Levy CLT.