Professional Documents
Culture Documents
Review of Probability Theory
Review of Probability Theory
Review of Probability Theory
#
Stat 504, Lecture 1 1
Review of
Discrete Probability
" $
! #
Stat 504, Lecture 1 2
2. P (not A) = 1 − P (A);
" $
! #
Stat 504, Lecture 1 3
" $
! #
Stat 504, Lecture 1 4
Random variables
A random variable is the outcome of an experiment
(i.e. a random process) expressed as a number. We
use capital letters near the end of the alphabet (X, Y ,
Z, etc.) to denote random variables. Random
variables are of two types: discrete and continuous.
Continuous random variables are described by
probability density functions (PDF). For example, a
normally distributed random variable has a
bell-shaped density function like this:
f (x) = P (X = x).
" $
! #
Stat 504, Lecture 1 6
.25
x
1 2 3 4 5
P
Note that x f (x) = 1 if the sum is taken over all
values of x having nonzero probability. In other
words, the sum of the heights of all the spikes must
equal one.
Joint distribution. Suppose that X1 , X2 , . . . , Xn are n
random variables, and let X be the entire vector
X = (X1 , X2 , . . . , Xn ).
f (x) = P (X = x)
" $
= P (X1 = x1 , X2 = x2 , . . . , Xn = xn ).
! #
Stat 504, Lecture 1 7
f (x) = P (X1 = x1 , X2 = x2 , . . . , Xn = xn )
= P (X1 = x1 ) P (X2 = x2 ) · · · P (Xn = xn )
= f (x1 ) f (x2 ) · · · f (xn )
Yn
= f (xi )
i=1
" $
! #
Stat 504, Lecture 1 8
" $
x
! #
Stat 504, Lecture 1 9
x f (x) g(x) = x3
1 .25 13 = 1
2 .25 23 = 8
5 .50 53 = 125
and
E(Y ) = E(X 3 ) = 1(.25) + 8(.25) + 125(.5) = 64.75.
If Y = g(X) = a + bX where a and b are constants,
then Y is said to be a linear function of X, and
E(Y ) = a + bE(X). An algebraic proof is
X
E(Y ) = yf (y)
y
X
= (a + bx)f (x)
x
X X
= af (x) + bxf (x)
x x
X X
= a f (x) + b xf (x)
x x
" $
= a · 1 + bE(X).
! #
Stat 504, Lecture 1 10
V (X) = E( (X − E(X))2 )
X
= (x − E(X))2 f (x).
x
" $
! #
Stat 504, Lecture 1 11
x f (x) x2
1 .25 12 = 1
2 .25 22 = 4
5 .50 52 = 25
First we calculate
E(X) = 1(.25) + 2(.25) + 5(.50) = 3.25 and
E(X 2 ) = 1(.25) + 4(.25) + 25(.50) = 13.75. Then
V (X) = 13.75 − (3.25)2 = 3.1875.
It can be shown that if a and b are constants, then
V (a + bX) = b2 V (X).
" $
! #
Stat 504, Lecture 1 12
V (X + Y ) = V (X) + V (Y ) if X, Y independent.
V (X + Y ) = V (X) + V (Y ) + 2 Cov(X, Y )
V (X + Y + Z) = V (X) + V (Y ) + V (Z)
+ 2 Cov(X, Y ) + 2 Cov(X, Z) + 2 Cov(Y, Z),
" $
mutually uncorrelated.