Professional Documents
Culture Documents
3 s2.0 B9780128104958000038 Main
3 s2.0 B9780128104958000038 Main
A = {stocks up Today}
B = {stocks up Yesterday}.
P (A ∩ B)
P (A|B) = , (2.1)
P (B)
when P (B) > 0.
The exclusion of the case P (B) = 0 has important ramifications later on.
If A and B are mutually exclusive events, then P (A|B) = 0. If A ⊆ B, then
P (A|B) = P (A)/P (B) ≥ P (A) with strict inequality unless P (B) = 1. If
B ⊆ A, then P (A|B) = 1. We will later have interest in the special case where
P (A|B) = P (A).
Note that P (·|B) is a probability measure that maps A → R+ . In particular,
we have:
Probability, Statistics and Econometrics. http://dx.doi.org/10.1016/B978-0-12-810495-8.00003-8
Copyright © 2017 Elsevier Inc. All rights reserved. 11
12 Probability, Statistics and Econometrics
Today
up down
up 53 25 78
Y’day
down 15 7 22
68 32 100
P (A) = P (A ∩ B) + P (A ∩ B c )
determines the marginals, and then definition (2.1) gives the conditionals. Thus
78
P (Up Y’day) =
100
22
P (Down Y’day) =
100
53/100 53
P (Up Today|Up Y’day) = = .
78/100 78
Proof. This follows from the definition of conditional probability and the law
of total probability:
The probability P (B) is frequently called the prior probability and P (A|B)
is called the likelihood, while P (B|A) is called the posterior. We next give
some examples.
Example 2.5. We next consider an example from economic theory, called the
sequential trading model. A stock price can take two possible values
VH with prob 1 − δ
V=
VL with prob δ,
where VL < VH and δ ∈ [0, 1]. This is the prior distribution on value. The in-
vestor is chosen randomly from two possible types
I with prob μ
T=
U with prob 1 − μ
The timeline is: first, a value is chosen, and then a type of investor is chosen,
and that investor carries out his strategy. The strategies of the investors are as
follows. The informed traders (I ), will buy if the value is high VH , and sell if
the value is low VL , provided the quoted prices lie in the interval [VL , VH ]. The
14 Probability, Statistics and Econometrics
uninformed traders (U ), buy or sell with probability 1/2. Suppose that a buy
order is received (but it is not known from whom), what does that tell us about
the value of the stock? Let A = {V = VL } and B = {buy order received}. We
can calculate P (B|A) directly from the knowledge of the traders strategies and
the distribution of trader types. That is
1
P (B|A) = (1 − μ)
2
because if the value is low the informed trader will not buy and the uninformed
trader will buy one half of the time. Likewise
1
P (B|Ac ) = (1 − μ) + μ
2
because when the value is high the informed trader will always buy. We want to
know P (A|B) as this tells us the updated value distribution. By Bayes rule, we
can calculate the updated distribution (posterior) of V
likelihood prior
posterior
P (B|A) P (A)
P (A|B) =
P (B)
P (buy|V = VL )P (V = VL )
=
P (buy|V = VL )P (V = VL ) + P (buy|V = VH )P (V = VH )
2 (1 − μ) × δ
1
=
(1 + μ(1 − 2δ))/2
1−μ
= ×δ
1 + μ(1 − 2δ)
≤δ
(μ + 1)
P (V = VH |buy) = P (Ac |B) = 1 − P (A|B) = (1 − δ) ≥ 1 − δ
(1 + μ(1 − 2δ)
The information that a buy order has been received is useful and increases our
valuation of the asset. On the other hand, if a sell order were received, this would
lower our valuation of the asset.
Example 2.6. Why most published research findings are false, Ioannidis (2005).
2.3 INDEPENDENCE
We next define the notion of independence, which is a central property in much
of statistics. This concerns a special case of conditional probability that makes
many calculations simpler.
Conditional Probability and Independence Chapter | 2 15
Definition 2.2. Independence. Suppose P (A), P (B) > 0, then A and B are
independent events if:
(1) P (A ∩ B) = P (A) · P (B)
(2) P (A|B) = P (A)
(3) P (B|A) = P (B)
We have
3 4 4 48 12 + 192 204 4 4
P (B) = · + · = = = <
51 52 51 52 51 · 52 51 · 52 52 51
So P (B) < P (B|A), i.e., A and B are not independent events. In this case
β(A, B) = 1/51 > 0 meaning there is positive dependence.
We next consider the more general case with more than two events.
Example 2.10. The infinite monkey theorem says that if one had an infinite
number of monkeys randomly tapping on a keyboard, with probability one, at
least one of them will produce the complete works of Shakespeare. If one has a
finite set of characters on the typewriter K and a finite length document n, then
the probability that any one monkey would type this document exactly is K −n .
If there are 47 keys on the standard typewriter, and 884,421 words, so perhaps
5 million characters, in which case the probability is so low that a given monkey
will produce the documents. Let Ai = {Monkey i nails it}. Then
P (Ai ) = K −n
This can be strengthened to say that with probability one an infinite number
of monkeys would produce the complete works of Shakespeare. Consider the
sets B1 = {1, 2, . . . , M}, B2 = {M + 1, M + 2, . . . , 2M}, etc. There was an Arts
18 Probability, Statistics and Econometrics
council grant that was commissioned to investigate the infinite monkey theo-
rem.1
We can have pairwise independence but not independence, i.e., pairwise in-
dependence is the weaker property.
Example 2.11. Suppose that S = {1, 2, 3, 4}, A = {1, 2}, B = {1, 3}, and C =
{1, 4}. Then P (A ∩ B ∩ C) = 1/4 but P (A) = P (B) = P (C) = 1/2 and P (A ∩
B) = P (A ∩ C) = P (B ∩ C) = 1/4.
Definition 2.5. Conditional Independence. Suppose that P (A), P (B), P (C) >
0, then A and B are independent events given C if either:
(1) P (A ∩ B|C) = P (A|C)P (B|C);
(2) P (C)P (A ∩ B ∩ C) = P (A ∩ C) · P (B ∩ C).
There are many examples where conditional independence holds but not in-
dependence. Likewise, the direction of dependence or association can change
1. In 2003, lecturers and students from the University of Plymouth MediaLab Arts course studied
the literary output of real monkeys. They left a computer keyboard in the enclosure of six Celebes
crested macaques in Paignton Zoo in Devon in England for a month, with a radio link to broadcast
the results on a website. The monkeys produced nothing but five total pages largely consisting of
the letter S, the lead male began by bashing the keyboard with a stone, and the monkeys continued
by urinating and defecating on it.
Conditional Probability and Independence Chapter | 2 19
This is called a Markov Chain. A special case of this is when pt is time invari-
ant as in our next example.
P1,0 = 1 − p + pP2,0 .
quadratic equation
P1,0 = 1 − p + pP1,0
2
,
which has solutions
1−p
P1,0 = 1 ; P1,0 = .
p
The first solution is relevant when p ≤ 1/2, while the second is relevant other-
wise. This says that for example even if you have a fifty fifty chance of success,
you will become bankrupt with probability one. This shows the advantage of the
principle “quit while you are ahead”.