Professional Documents
Culture Documents
Probability Convergence PDF
Probability Convergence PDF
9 Convergence in probability
The idea is to extricate a simple deterministic component out of a random situation. This is
typically possible when a large number of random effects cancel each other out, so some limit
is involved. The general situation, then, is the following: given a sequence of random variables,
Y1 , Y2 , . . ., we want to show that, when n is large, Yn is approximately f (n) for some simple
deterministic function f (n). The meaning of approximately is what we now make clear.
Example 9.1. Toss a fair coin n times, independently. Let Rn be the longest run of heads,
i.e., the longest sequence of consecutive tosses of Heads. For example, if n = 15 and the tosses
come out
HHTTHHHTHTHTHHH.
in probability. This means that, to a first approximation, one should expect about 20 consecutive
heads somewhere in a million tosses.
To solve a problem such as this, we need to find upper bounds on probabilities that Rn is
large and that it is small, i.e., on P (Rn k) and P (Rn k), for appropriately chosen k. Now,
for arbitrary k,
For the lower bound, divide the string of size n into disjoint blocks of size k. There is nk
such blocks (if n is not divisible by k, simply throw away the leftover smaller block at the end).
Then Rn k as soon as one of the blocks consists of Heads only, and different blocks are
independendent. Therefore,
n
1 k 1 jnk
P (Rn < k) 1 k exp k ,
2 2 k
9 CONVERGENCE IN PROBABILITY 112
as
Rn Rn Rn
P
1
= P 1 + or 1
log2 n log2 n log2 n
Rn Rn
= P 1+ +P 1
log2 n log2 n
= P (Rn (1 + ) log2 n) + P (Rn (1 ) log2 n) .
A little bit of fussing in the proof comes from the fact that (1 e) log2 n are not integers. This
common in problems such as this. To prove (1), we plug k = (1 + ) log2 n into the upper
bound to get
1
P (Rn (1 + ) log2 n) n
2(1+) log2 n1
2
= n
n1+
2
= 0
n
as n . One the other hand, to prove (2) we need to plug k = (1 ) log2 n + 1 into the
lower bound,
2
P (|X | k) ,
k2
for any k > 0. We proved this inequality in the previous chapter, and we will use it to prove the
next theorem.
EYn a,
Var(Yn ) 0,
as n . Then
Yn a,
as n , in probability.
then
Sn = X1 + . . . + Xn ,
where Xi are random variables with finite expectation and variance. Then, without any inde-
pendence assumption,
ESn = EX1 + . . . + EXn
and
n
X X
E(Sn2 ) = EXi2 + E(Xi Xj ),
i=1 i6=j
Xn X
Var(Sn ) = Var(Xi ) + Cov(Xi , Xj ).
i=1 i6=j
9 CONVERGENCE IN PROBABILITY 114
Continuing with the review, lets reformulate and reprove the most famous convergence in prob-
ability theorem. We will use the common abbreviation i. i. d. for independent identically dis-
tributed random variables.
Theorem 9.2. Weak law of large numbers. Let X, X1 , X2 , . . . be i. i. d. random variables with
with EX1 = and Var(X1 ) = 2 < . Let Sn = X1 + . . . + Xn . Then, as n ,
Sn
n
in probability.
Sn
Proof. Let Yn = n . We have EYn = , and
1 1 2 2
Var(Yn ) = Var(S n ) = n = .
n2 n2 n
Thus we can simply apply the previous theorem.
Example 9.2. We analyze a typical investment (the accepted euphemism for gambling on
financial markets) problem. Assume you have two investment choices at the beginning of each
year:
a risky stock which increases your investment by 50% with probability 0.8 and wipes it
away with probability 0.2.
Putting an amount s in the bond, then, gives you 1.06s after a year. The same amount in the
stock gives you 1.5s with probability 0.8 and 0 with probabiliy 0.2; note that the expected value
is 0.8 1.5s = 1.2s > 1.06s. We will assume year-to-year independence of the stocks return.
We will try to maximize the return to our investment by hedging. That is, we invest, at
the beginning of each year, a fixed proportion x of our current capital into the stock and the
remaining proportion 1 x into the bond. We collect the resulting capital at the end of the
year, which is simultaneously the beginning of next year, and reinvest with the same proportion
x. Assume that our initial capital is x0 .
9 CONVERGENCE IN PROBABILITY 115
It is important to note that the expected value of the capital at the end of the year is
maximized when x = 1, but using this strategy you will eventually lose everything. Let Xn be
your capital at the end of year n. Define the average growth rate of your investment as
1 Xn
= lim log ,
n n x0
so that
Xn x0 en .
We will express in terms of x; in particular, we will show it is a nonrandom quantity.
Let Ii = I{stock goes up in year i} . These are independent indicators with EIi = 0.8.
where Sn = I1 + . . . + In . Therefore,
1 Xn Sn Sn
log = log(1.06 + 0.44x) + 1 log(1.06(1 x))
n x0 n n
0.8 log(1.06 + .44x) + 0.2 log(1.06(1 x)),
Example 9.3. Distribute n balls independently at random into n boxes. Let Nn be the number
of empty boxes. Show that n1 Nn converges in probability and identify the limit.
Note that
Nn = I1 + . . . + In ,
where Ii = I{ith box is empty} , but you cannot use the weak law of large numbers as Ii are not
independent. Nevertheless,
n1 n 1 n
EIi = = 1 ,
n n
and so
1 n
ENn = n 1 .
n
9 CONVERGENCE IN PROBABILITY 116
Moreover, X
E(Nn2 ) = ENn + E(Ii Ij )
i6=j
with n
n2
E(Ii Ij ) = P (box i and j are both empty) = ,
n
so that
1 n 2 n 1 2n
Var(Nn ) = E(Nn2 ) 2
(ENn ) = n 1 + n(n 1) 1 2
n 1 .
n n n
as n . Therefore
Nn
Yn = e1 ,
n
as n , in probability.
Problems
1. Assume that n married couples (amounting to 2n people) are seated at random on 2n seats
around a table. Let T be the number of couples that sit together. Determine ET and Var(T ).
2. There are n birds that sit in a row on a wire. Each bird looks left or right with equal
probability. Let N be the number of birds not seen by any neighboring bird. Determine, with
proof, the constant c so that, as n , n1 N c in probability.
3. Recall the coupon collector problem: sample from n cards, with replacement, indefinitely,
and let N be the number of cards you need to get each of n different cards are represented. Find
a sequence an so that, as n , N/an converges to 1 in probability.
4. Kings and Lakers are playing a best of seven playoff series, which means they play until
one team wins four games. Assume Kings win every game independently with probability p.
9 CONVERGENCE IN PROBABILITY 117
(There is no difference between home and away games.) Let N be the number of games played.
Compute EN and Var(N ).
5. An urn contains n red and m black balls. Pull balls from the urn one by one without
replacement. Let X be the number of red balls you pull before any black one, and Y the
number of red balls between the first and the second black one. Compute EX and EY .
Solutions to problems
1. Let Ii be the indicator of the event that the ith couple sits together. Then T = I1 + + In .
Moreover,
2 22 (2n 3)! 4
EIi = , E(Ii Ij ) = = ,
2n 1 (2n 1)! (2n 1)(2n 2)
for any i and j 6= i. Thus
2n
ET =
2n 1
and
4 4n
E(T 2 ) = ET + n(n 1) = ,
(2n 1)(2n 2) 2n 1
so
4n 4n2 4n(n 1)
Var(T ) = 2
= .
2n 1 (2n 1) (2n 1)2
1
2. Let Ii indicate the event that bird i is not seen by any other bird. Then EIi is 2 if i = 1 or
i = n and 41 otherwise. It follows that
n2 n+2
EN = 1 + = .
4 4
Furthermore Ii and Ij are independent if |i j| 3 (two birds that have two or more birds
between them are observed independently). Thus Cov(Ii , Ij ) = 0 if |i j| 3. As Ii and Ij are
indicators, Cov(Ii , Ij ) 1 for any i and j. For the same reason Var(Ii ) 1. Therefore
X X
Var(N ) = Var(Ii ) + Cov(Ii , Ij ) n + 4n = 5n.
i i6=j
1 1 1 1
Clearly, if M = nN, then EM = n EN 4 and Var(M ) = n2 Var(N ) 0. It follows that
c = 14 .
3. Let Ni be the number of coupons needed to get i different coupons after having i 1 different
ones. Then N = N1 + . . . + Nn , and Ni are independent Geometric with success probability
9 CONVERGENCE IN PROBABILITY 118
ni+1
n . So
n n(i 1)
ENi = , Var(Ni ) = ,
ni+1 (n i + 1)2
and therefore
1 1
EN = n 1 + + . . . + ,
2 n
n 2
X n(i 1) 2 1 1 2
Var(N ) = n 1 + + . . . + n < 2n2 .
(n i + 1)2 22 n2 6
i=1
If an = n log2 n, then
1 1
EN 1, EN 0,
an a2n
as n , so that
1
N 1
an
in probability.
4. Let Ii be the indicator of the event that the ith game is played. Then EI1 = EI2 = EI3 =
EI4 = 1,
EI5 = 1 p4 (1 p)4 ,
EI6 = 1 p5 5p4 (1 p) 5p(1 p)4 (1 p)5 ,
6 3
EI7 = p (1 p)3 .
3
Add the seven expectations to get EN . Now to compute E(N 2 ) we use the fact that, if i > j,
Ii Ij = Ii , so that E(Ii Ij ) = EIi . So
X X X X 7
X
EN 2 = EIi + 2 E(Ii Ij ) = EIi + 2 (i 1)EIi = (2i 1)EIi ,
i i>j i i i=1
and the final result can be obtained by plugging in EIi and finally by the standard formula
5. Imagine the balls ordered in a row, and the ordering specifies the sequence in which they are
pulled. Let Ii be the indicator of the event that the ith red ball is pulled before any black ones.
1
Then EIi = m+1 , simply the probability that in a random ordering of the ith red balls and all
n
m black ones, the red comes first. As X = I1 + . . . + In , EX = m+1 .
Now let Ji be the indicator of the event that the ith red ball is pulled between the first and
the second black one. Then EJi is the probability that the red ball is second in the ordering of
m + 1 balls as above, so EJi = EIi , and EY = EX.