Professional Documents
Culture Documents
Moments and Moment Generating Functions
Moments and Moment Generating Functions
Moments
Introduction
Figure 1 Two PDFs with mean 2 and variance 12. The light curve is Normal and
symmetric; the dark curve is Log-Normal and right-skewed.
The Normal curve is symmetric about 2, so its mean, median, and mode are all 2. In
s
contrast, the Log-Normal is heavily skewed to the right; this means its right tail is very
long compared to its left tail. It has mean 2, but median 1 and mode 0.25. From the
mean and variance alone, we would not be able to capture the difference between the
asymmetry of the Log-Normal and the symmetry of the Normal. We will learn that a
standard measure of the asymmetry of a distribution is based on the third moment.
Summaries of a distribution (contd.)
Figure 2 Left: N(0, 1) PDF (light) and a scaled t3 PDF (dark). Both have mean 0
and variance 1, but the latter has a sharper peak and heavier tails. Right: magnified
version of right tail behavior.
The previous example considered difference between symmetric and asymmetric
s the same mean and variance can also
distributions, but symmetric distributions with
look very different. The left plot in above figure shows two symmetric PDFs with
mean 0 and variance 1. The dark curve has a sharper peak and heavier tails than the
light curve. The tail behavior is magnified in the right plot, where it is easy to see that
the dark curve decays to 0 much more gradually, making outcomes far out in the tail
much more likely than for the light curve. A standard measure of the peakedness of a
distribution and the heaviness of its tails is based on the fourth moment.
Interpreting moments
Proposition 1.
Odd central moments of a symmetric distribution: Let X be symmetric
about its mean µ. Then for any odd number m, the mth central moment
E (X − µ)m is 0 if it exists.
Proof. Since X − µ has the same distribution as µ − X , they have the
same mth moment (if it exists):
E (X − µ)m = E (µ − X )m
∞
X tn
M(t) = E (X n )
n=0
n!
Theorem 3.
MGF determines the distribution: The MGF of a random variable
determines its distribution: if two r.v.s have the same MGF, they must
have the same distribution. In fact, if there is even a tiny interval (−a, a)
containing 0 on which the MGFs are equal, then the r.v.s must have the
same distribution.
Theorem 4.
MGF of a sum of independent r.v.s: If X and Y are independent, then the
MGF of X + Y is the product of the individual MGFs:
Using the last theorem, we can easily get the MGFs of the Binomial and
Negative Binomial.
I Binomial MGF The MGF of a Bern(p) r.v. is pe t + q, so the MGF
of a Bin(n, p) r.v. is
p
M(t) = ( )r
1 − qe t
for qe t < 1.
Moment generating functions (contd.)
Proposition 2.
MGF of location-scale transformation: If X has MGF M(t), then the MGF of a + bX is
On the other hand, we know that E (X n ) is the coefficient of the term involving
t n in the Taylor expansion of M(t):
∞
X tn
M(t) = E (X n )
n=0
n!
I Thus we can match coefficients to conclude that E (X n ) = n! for all n. We not
only did not have to do a LOTUS integral, but also we did not, for example,
have to take 10 derivatives to get the 10th moment—we got the moments all at
once.
I To find the moments of Y ∼ Expo(λ), use a scale transformation: we can
express Y = X /λ where X ∼ Expo(1). Therefore Y n = X n /λn and
n!
E (Y n ) =
λn
I We have found the mean and variance of Y
1
E (Y ) =
λ
2 1 1
Var (Y ) = E (Y 2 ) − (EY )2 = − 2 = 2
λ2 λ λ