Professional Documents
Culture Documents
Probabilistic Methods in Engineering: Dr. Horst Hohberger
Probabilistic Methods in Engineering: Dr. Horst Hohberger
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 1 / 355
Elements of Probability Theory
Part I
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 2 / 355
Elements of Probability Theory
Joint Distributions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 3 / 355
Elements of Probability Theory Introduction to Probability and Counting
Joint Distributions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 4 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 5 / 355
Elements of Probability Theory Introduction to Probability and Counting
In the 16th century, the Italian mathematician Cardano, who was a heavy
gambler, attempted to use mathematics to describe the outcome of
games. He hit upon the following definition, which is really a procedure for
calculating probabilities:
1.1.1. Definition. Let A be a random outcome (random event) of an
experiment (game) that may proceed in various ways. Assume each of
these ways is equally likely. Then the probability of the outcome A is
number of ways leading to outcome A
P(A) =
number of ways the experiment can proceed.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 6 / 355
Elements of Probability Theory Introduction to Probability and Counting
1.1.2. Examples.
1. The experiment consists of flipping a coin. We are interested in the
probability of the coin landing heads up. The experiment can proceed
in two ways: the coin land heads up or tails up. We assume each
event is equally likely, so the classical definition gives
1
P[coin lands heads up] = = 0.5
2
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 7 / 355
Elements of Probability Theory Introduction to Probability and Counting
1.1.3. Examples.
2. The experiment consists of rolling two 6-sided dice and summing the
results, so the possible outcomes are the numbers S = 2, 3, . . . , 12.
We are interested in the outcome S = 3. Each die will give results 1,
2, 3, 4, 5 or 6. We assume that each result is equally likely. There are
two ways we can get the outcome S = 3: either the first die’s result is
1 and the second die’s result is 2, or the first die gives 2 and the
second die gives 1. In total, the experiment can proceed in 6 × 6 = 36
different ways. Hence
2 1
P[3] = = = 0.056
36 18
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 8 / 355
Elements of Probability Theory Introduction to Probability and Counting
Historical Notes
Cardano’s work although published, received little attention, and 100 years
later, in the middle of the 17th century, the two French mathematicians
Fermat and Pascal rediscovered his principles, also by considering games.
They discussed how to divide the jackpot if a game in progress is
interrupted. Imagine that Fermat and Pascal are playing a simple game,
whereby a coin is repeatedly tossed. Fermat wins as soon as the coin has
turned up heads six times, Pascal wins as soon as the coin has turned up
tails six times.
There are 24 gold pieces in the pot. Now the game is interrupted when
the coin has already turned up 5 tails and 4 heads. How to divide the pot?
Fermat can only win if the coin turns up heads two times in a row, and we
can calculate that he has a 1/4 chance of this happening. Therefore, he
receives 1/4 of the pot (6 gold pieces), while Pascal receives 3/4 (18 gold
pieces).
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 9 / 355
Elements of Probability Theory Introduction to Probability and Counting
Tree Diagrams
1 Toss st
ee eeeeeeeYYYYYYYYtails
YYYYYY
eee
heads
eeeeee YYYY
2nd Toss
R 2nd Toss
R
l RRRRtails l RRRRtails
llll
heads
l RRRR llll
heads
l RRRR
llll R llll R
2 heads 1 head, 1 tail 1 head, 1 tail 2 tails
After two tosses, there are four possible outcomes. Only one of the four
outcomes involves two heads, so the probability of this happening is 1/4.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 10 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 11 / 355
Elements of Probability Theory Introduction to Probability and Counting
Permutations in Calculus
There is a small difference in the use of the term “permutation” between
analysts and combinatorists; we first repeat the usage that is familiar from
calculus/linear algebra.
Permutations in Calculus
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 13 / 355
Elements of Probability Theory Introduction to Probability and Counting
Permutations in Calculus
1.1.7. Example. Consider the set of 3 objects, {1, 2, 3}. Then one
permutation is given by the identity map,
µ ¶
1 2 3
π0 =
1 2 3
while other permutations are given by
µ ¶ µ ¶ µ ¶ µ ¶ µ ¶
1 2 3 1 2 3 1 2 3 1 2 3 1 2 3
, , , , .
3 1 2 2 3 1 3 2 1 2 1 3 1 3 2
1.1.8. Notation. Instead of writing out the map as above, we often simply
give the ordered n − tuple (π(1), . . . , π(n)). For example, instead of
µ ¶
1 2 3
we might write (3, 1, 2).
3 1 2
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 14 / 355
Elements of Probability Theory Introduction to Probability and Counting
Permutations in Calculus
We have previously encountered permutations in the definition of the
determinant of an n × n matrix using the Leibniz formula; there one sums
over all permutations of the index set {1, . . . , n} and also includes the sign
of the permutation, a concept we won’t elaborate on here.
Proof.
We consider the number of possible values of π(k), k = 1, . . . , n. The first
value, π(1) can be any of the n elements of {1, . . . , n}. The second value
can be any element of {1, . . . , n} \ {π(1)}. Hence there are n possible
values of π(1), but only n − 1 possible values of π(2). In general, there are
n − k + 1 possible values for π(k), so in total, there are
n · (n − 1) · · · (n − n + 1) = n!
Permutations in Combinatorics
In combinatorics, permutations are regarded as arrangements in a definite
order. Clearly, the ordered tuple (π(1), . . . , π(n)) is an arrangement of
(1, . . . , n) is a definite order, i.e., the calculus-based definition realizes this
goal.
However, in combinatorics one is often interested in first selecting r ≤ n
objects from {1, . . . , n} and then arranging these objects in a definite
order. We realize this by the following, more general definition:
π : {1, . . . , r } → {1, . . . , n}
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 16 / 355
Elements of Probability Theory Introduction to Probability and Counting
Permutations in Combinatorics
1.1.12. Notation. We again write π as
µ ¶
1 2 ... r
or (π(1), . . . , π(r )),
π(1) π(2) . . . π(r )
where π(k) ∈ {1, . . . , n}, k = 1, . . . , r .
(1, 2), (2, 1), (1, 3), (3, 1), (2, 3), (3, 2).
Combinations
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 18 / 355
Elements of Probability Theory Introduction to Probability and Counting
Combinations
1.1.17. Theorem. There are
µ ¶
n n!
=
r r !(n − r )!
combinations of r elements from {1, . . . , n}.
Proof.
We first consider permutations of r objects from {1, . . . , n}. Every
permutation gives us a combination, by identifying
Obviously, more than one permutation will give us the same combination.
Note that any permutation of (π(1), . . . , π(r )) will be a permutation of r
objects from {1, . . . , n}. Furthermore, any permutation of (π(1), . . . , π(r ))
will yield the same combination. For each tuple (π(1), . . . , π(r )), there are
r ! permutations of (π(1), . . . , π(r )), so we need to divide the total number
of permutations of r objects from {1, . . . , n} by r !.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 19 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 20 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 21 / 355
Elements of Probability Theory Introduction to Probability and Counting
n!
n1 !n2 ! . . . nk !
possible ways of dividing N into k sets A1 , . . . , Ak , where
S
k Tk
◮ N = Ai , Ai = ∅
i=1 i=1
◮ |Ai | = ni , i = 1, . . . , k.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 22 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 23 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 24 / 355
Elements of Probability Theory Introduction to Probability and Counting
Events
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 26 / 355
Elements of Probability Theory Introduction to Probability and Counting
Example
1.1.25. Example. We roll a four-sided die 10 times. What is the probability
of obtaining 5 ones, 3 twos, 1 three and 1 four?
There are 410 = 1048576 possibilities for the 10-tuple of results of the die
rolls, corresponding to that many sample points in S = N10 that
correspond to physical results. The event A consists of all ordered
10-tuples containing 5 ones, 3 twos, 1 three and 1 four. There are
10!
= 5040
5!3!1!1!
possible ways of obtaining 5 ones, 3 twos, 1 three and 1 four, so there are
that many elements in A. The probability is
5040
≈ 0.00481 ≈ 0.5%.
1048576
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 27 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 28 / 355
Elements of Probability Theory Introduction to Probability and Counting
Counting Mountains
We will now look at a non-trivial example of counting. Consider the
following problem: We wish to draw “mountains” using upstrokes and
downstrokes, like this:
◮ One upstroke, one downstroke:
B
||| BBB
|| B
6
©©© 666
© 66
©©© 66
B B ©© 66
||| BBB ||| BBB ©© 66
|| B|| B ©© 6
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 29 / 355
Elements of Probability Theory Introduction to Probability and Counting
B 6 6
©©© BBB ||| 666 ©©© 666
© B|| 66 © 66
©©© 66 ©©© 66
© 66 © 66
©©© 66 ©©© 66 ||BBB
6 6||| BB
©© ©©
4
444
44
44
44 6
44
©©© 666
44 © 66
44 ©© 66
44 BB ©©© 66
44 |
| BB ©© 66
4 | 6
|| B©©
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 30 / 355
Elements of Probability Theory Introduction to Probability and Counting
Counting Mountains
The mountains must each consist of n upstrokes and n downstrokes, and
thre may be no “valleys,” i.e., no stroke may lie under the starting point.
That means that the following constructions are not allowed:
BB B |66
BB ||| BBB ||| || 66
B|| B|| || 66
66
66
66 ||
6|||
The question is: for given n, how many such mountains using n up- and n
downstrokes can we draw?
The answer is related to many diverse combinatorial problems.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 31 / 355
Elements of Probability Theory Introduction to Probability and Counting
Counting Mountains
6
©©© 666
© 66
©©© 66
© 66
©©© 66 ||BBB
© 6||| BB
©
as W(3 ↑, 3 ↓) =↑↑↓↓↑↓.
However, not all such words do not lead to admissible mountains. For
example, ↑↓↓↑↓↑ does not give a mountain.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 32 / 355
Elements of Probability Theory Introduction to Probability and Counting
A proof will be given in the recitation class. If you think that you are good
at counting, I challenge you to find a proof by yourselves before the class!
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 33 / 355
Elements of Probability Theory Introduction to Probability and Counting
Towards the end of the 17th century, the French mathematician de Moivre
had to flee to England, as he was a protestant (Huguenot), who were
being persecuted in France at the time. However, in England he couldn’t
get a job, since he was French. (The English did not like or trust the
French.) So he ended up spending a lot of time in coffee houses.
There he earned money by helping gamblers resolve disputes (such as how
to divide the pot in an interrupted game). His experiences led him to the
concept of independence of events - a coin does not remember its last
result, and will always have the same chance of turning heads up, no
matter how often it has done so in the past. We will explore this essential
concept in more detail later.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 34 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 35 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 36 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 37 / 355
Elements of Probability Theory Introduction to Probability and Counting
1.1.29. Remark. You should be careful in making deductions from the law
of large numbers. For example, if a fair coin is tossed four times, it is more
likely to obtain a 3-1 split (3 heads and one tail or 3 tails and one head)
than it is to obtain a 2-2 split (2 heads and 2 tails). Verify this for yourself
by using the classical definition of probability!
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 38 / 355
Elements of Probability Theory Introduction to Probability and Counting
What is Probability?
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 40 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 41 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 42 / 355
Elements of Probability Theory Introduction to Probability and Counting
Now that we have a good grasp of the concept of probability, we can use
tree diagrams more generally. Consider the case of an unfair coin, which
turns heads up 60% of the time. We toss the coin twice:
eeeeYYYYYYYYY0.4
eeeee eeeeee
0.6 YYYYYY
YYY
eee
heads
R tails
R
0.6lllll RRRR0.4 0.6lllll RRRR0.4
l RRRR l RRRR
llll R llll R
2 heads head, tail tail, head 2 tails
P[2 heads] = 0.36 P[head,tail] = 0.24 P[tail,head] = 0.24 P[2 tail] = 0.16
Probabilities are multiplied along each branch to give the final probability,
and summed across sub-branches to give unity.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 43 / 355
Elements of Probability Theory Introduction to Probability and Counting
Conditional Probability
We now want to study the effect of information on probability. In
particular, two events may be related, such that the occurrence of one
event influences the probability that the other event occurs. For example,
event B may have a 50% chance of occurring, but if one also knows that
event A has previously occurred, then event B might have a higher or
lower chance of occurring.
1.1.31. Example. There is a test for the gender of an unborn child called
“starch gel electrophoresis” (this type of test is of course not done in
China!). It detects the presence of a protein zone called the pregnancy
zone. The following statistical information is known:
◮ 43% of all pregnant women have the pregnancy zone.
◮ 51% of all children born are male.
◮ 17% of all children are male and their mothers have the pregnancy
zone.
Given that the zone is present, what is the probability that the child is
male?
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 44 / 355
Elements of Probability Theory Introduction to Probability and Counting
Conditional Probability
We now use a classical probability tree,
All pregnant
fXX
Women
ffff XXXXX
fffff XXXXXP[zone present] = 0.43
fffff XXX
Zone Absent Zone present
P
nn PPPP
nnnnn PPP P[male | zone present]
n
Female Male
P[male ∩ zone present] = 0.17
Here P[male | zone present] means the probability that the child is male,
given that the zone is present.
Since we multiply probabilities along branches of probability trees, it is
clear that
P[male ∩ zone present] 0.17
P[male | zone present] = = ≈ 0.40
P[zone present] 0.43
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 45 / 355
Elements of Probability Theory Introduction to Probability and Counting
Conditional Probability
The previous example motivates the general definition:
1.1.32. Definition. Let A, B be events and P(A) ̸= 0. Then we define the
conditional probability
P[A ∩ B]
P[B | A] := .
P(A)
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 46 / 355
Elements of Probability Theory Introduction to Probability and Counting
You are participating in a game show to win 10,000,000 RMB. The game
master [Monty Hall] presents you with three closed doors. Behind one of
the doors is the prize, behind the other two doors there is nothing. If you
open the correct door, you will receive the money, if you open one of the
other two doors you will not get anything.
Before opening any of the three doors, you can announce which door you
intend to open. Obviously, at least one of the other two doors does not
hide the money. The game master opens this (empty) door. You are then
given the option of either
◮ sticking with your choice or
◮ switching to the other closed door.
What do you do and does it make a difference?
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 47 / 355
Elements of Probability Theory Introduction to Probability and Counting
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 48 / 355
Elements of Probability Theory Introduction to Probability and Counting
Independence of Events
If one event does not influence another, then we say that the two events
are independent. Mathematically, we express this in the following way.
1.1.33. Definition. Let A, B be two events. We say that A and B are
independent if
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 49 / 355
Elements of Probability Theory Introduction to Probability and Counting
Independence of Events
1.1.34. Example. The birthdays (day and month) of a group of people are
generally assumed to be independent. Disregarding leap years, any person
is assumed to have a 1/365 chance of being born on a given day. (Do you
think that this is a reasonable assumption?) How many people should a
group have so that there is a better than even chance of two people in the
group having the same birthday?
We consider the complementary problem and start with a single person in
the group. If we add a second person, there is a 364/365 chance of them
not sharing a birthday. Adding a third person, for no two people to share a
birthday, this person must have his birthday on one of the other 363 days
of the year, so there is now a
364 363
365 365
chance of no two people in the group sharing a birthday.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 50 / 355
Elements of Probability Theory Introduction to Probability and Counting
Independence of Events
Continuing this argument, in a group of n ≥ 2 people there is a
Y
n
366 − k 1 364!
=
365 365 n−1 (365 − n)!
k=2
chance of no two people having the same birthday. It turns out that for
n = 23 this number is less than 0.5, so the probability of two people
having the same birthday is > 0.5.
This statement has been verified empirically; in a soccer match there are
2 × 11 players + 1 referee on the pitch. On any given playing day in the
Premier Division of the English league, about half the games should
feature two participants with the same birthday.
Refer to the article Coincidences: The truth is out there, Teaching
Statistics, Vol. 1, No. 1, 1998 (uploaded to the Resources section on
SAKAI).
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 51 / 355
Elements of Probability Theory Introduction to Probability and Counting
Total Probability
Imagine you conduct a 2-stage experiment with two possible outcomes A
and B but three intermediate stages F , G and H. The probability tree for
this experiment looks like this:
P[F ] ccccccccccc[[[[[[[[[[[P[H]
ccccccc [[[[[[[[
ccccccc P[G ] [[[[[[
F G H
P[A|F ]qqqMMMP[B|F
M ] P[A|G ]qqqMMMP[B|G
M ] P[A|H]qqqMMMP[B|H]
M
q MMM q MMM q MMM
qqq qqq qqq
A B A B A B
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 52 / 355
Elements of Probability Theory Introduction to Probability and Counting
Total Probability
X
n
P[B] = P[B | Aj ] · P[Aj ].
j=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 53 / 355
Elements of Probability Theory Introduction to Probability and Counting
Bayes’s Theorem
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 54 / 355
Elements of Probability Theory Introduction to Probability and Counting
Bayes’s Theorem
1.1.37. Example. Blood type distribution in US:
◮ Type A: 41%
◮ Type B: 9%
◮ Type AB: 4%
◮ Type O: 46%
Measurements statistics:
◮ P[type A registered | true type O] = 4%
◮ P[type A registered | true type A] = 88%
◮ P[type A registered | true type B] = 4%
◮ P[type A registered | true type AB] = 10%
A test registers type A; what is the probability that this is correct?
Wanted: P[true type A | type A registered]
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 55 / 355
Elements of Probability Theory Introduction to Probability and Counting
Bayes’s Theorem
By Bayes’s Theorem,
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 56 / 355
Elements of Probability Theory Discrete Random Variables
Joint Distributions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 57 / 355
Elements of Probability Theory Discrete Random Variables
1.2.2. Examples.
1. The integer numbers are countable.
2. The rational numbers are countable.
3. The real numbers in the interval [0, 1] are uncountable.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 58 / 355
Elements of Probability Theory Discrete Random Variables
Please note:
◮ The map X : S → Ω alone is just a variable.
◮ If Ω is countable, we call X a discrete variable, otherwise a
continuous variable.
◮ X together with fX (i.e., the pair (X , fX )) is a discrete or continuous
random variable.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 61 / 355
Elements of Probability Theory Discrete Random Variables
Cumulative distribution
In the previous example, we might be interested in the probability that not
more than 10 out of 25 students have blood type A. The random variable
would be
X : S → Ω = {0, 1, 2, . . . , 25}, (n1 , n2 , n3 , n4 ) 7→ x = n1 .
with a certain density function fX . We would hence determine
P[X ≤ 10] = P[X = 0] + P[X = 1] + · · · + P[X = 10].
This is known as the cumulative probability, and we in fact define the
so-called cumulative distribution function of a random variable by
F (x) = P[X ≤ x].
For a discrete random variable,
X X
F (x) = P[X = y ] = fX (y ).
y ≤x y ≤x
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 62 / 355
Elements of Probability Theory Discrete Random Variables
Expectation
Consider the rolling of a fair six-sided die. We are interested in the average
or expected value of the result. Since each result (numbers 1,2,3,4,5,6)
occurs with probability 1/6, we take the weighted sum:
1 1 1 1 1 1
· 1 + · 2 + · 3 + · 4 + · 5 + · 6 = 3.5.
6 6 6 6 6 6
The average result of a die roll is then 3.5, even though this result itself
can never occur.
1.2.5. Definition. Let (X , fX ) be a discrete random variable. Then the
expected value of X is
X
E[X ] = x · fX (x).
x∈Ω
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 63 / 355
Elements of Probability Theory Discrete Random Variables
Expectation
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 64 / 355
Elements of Probability Theory Discrete Random Variables
3. Let X , Y be two random variables that are independent (we will later
give a precise definition). Then
E[X + Y ] = E[X ] + E[Y ].
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 65 / 355
Elements of Probability Theory Discrete Random Variables
Variance
The expectation of a random variable is its mean, the point about which
the values of the random variable fluctuate. What the mean does not
describe is the degree of this fluctuation.
1.2.8. Example. We roll three fair six-sided dice and sum the results. The
result of each die roll is a random variable X , Y and Z , respectively. The
random variable is X + Y + Z and its expected value is
E[X + Y + Z ] = E[X ] + E[Y ] + E[Z ] = 3.5 + 3.5 + 3.5 = 10.5
We can also roll a single twenty-sided (icosahedral) die. Then the
expected value of the result is
X
20
i 1 20 · 21
E[W ] = = = 10.5
20 20 2
i=1
Although both W and X + Y + Z have the same mean, they assume a
totally different range of values and we would also find that
P[X + Y + Z = 9] ̸= P[W = 9].
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 66 / 355
Elements of Probability Theory Discrete Random Variables
Variance
We therefore want to measure not only the mean of a random variable,
but also its expected deviation from the mean. The deviation from the
mean is given by
X − E [X ],
but we are interested in its absolute size, so we square it and then take the
expected value.
1.2.9. Definition. Let X be a random variable with expectation E[X ]. Then
the variance of X is defined as
£ ¤
Var X = E (X − E[X ])2 .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 67 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 68 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 69 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 70 / 355
Elements of Probability Theory Discrete Random Variables
Geometric Properties.
1. The experiment consists of a series of trials. The outcome of each
trial can be classed as being either a “success” (s) or a “failure” (f ).
A trial with this property is called a Bernoulli trial.
2. The trials are identical and independent in the sense that the
outcome of one trial has no effect on the outcome of any other. The
probability of success, p, remains the same for each trial.
3. The random variable X denotes the number of trials needed to obtain
the first success.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 71 / 355
Elements of Probability Theory Discrete Random Variables
X : S → Ω = N \ {0}
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 73 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 74 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 75 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 76 / 355
Elements of Probability Theory Discrete Random Variables
pe t
mX : (−∞, − ln q) → R, mX (t) =
1 − qe t
where q = 1 − p.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 77 / 355
Elements of Probability Theory Discrete Random Variables
Proof.
We have fX (x) = q x−1 p for x ∈ N \ {0}. Then
∞
X ∞
pX t x
mX (t) = E[e tX ] = e tx q x−1 p = (qe )
q
x=1 x=1
This is a geometric series which converges for |qe t | = qe t < 1, i.e., for
t < − ln q. For such t, the limit is given by
p ³X t x ´ p³ 1 ´
∞ ∞
pX t x
mX (t) = (qe ) = (qe ) − 1 = − 1
q q q 1 − qe t
x=1 x=0
p qe t pe t
= = .
q 1 − qe t 1 − qe t
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 78 / 355
Elements of Probability Theory Discrete Random Variables
Proof.
We use the moment-generating function to calculate the expectation value:
¯ ¯
d ¯¯ d ¯¯ pe t
E [X ] = ¯ m X (t) = ¯
dt t=0 dt t=0 1 − qe t
¯
pe (1 − qe ) + pe t qe t ¯¯
t t p 1
= ¯ = = .
(1 − qe )t 2
t=0 (1 − q)2 p
The proof for the variance is similar and is left to the reader.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 79 / 355
Elements of Probability Theory Discrete Random Variables
We will now look at various further discrete distributions, starting with the
binomial distribution.
Binomial Properties.
1. The experiment consists of a fixed number n of Bernoulli trials.
2. The trials are identical and independent. The probability of success,
p, remains the same for each trial.
3. The random variable X denotes the number of successes in the n
trials.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 80 / 355
Elements of Probability Theory Discrete Random Variables
X : S → Ω = {0, 1, 2, . . . , n}
2. E[X ] = np.
3. Var X = npq.
The proof of this theorem is left as an exercise. We will also later be very
interested in the cumulative distribution F , F (x) = P[X ≤ x]. There is no
simple way of evaluating the sums involved, so the values have been
tabulated (Table I of Appendix A in the textbook). The table gives the
values of
⌊t⌋ µ ¶
X n x
F (t) = p (1 − p)n−x ,
x
x=0
where ⌊t⌋ is the greatest integer less than or equal to t.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 82 / 355
Elements of Probability Theory Discrete Random Variables
Pascal Properties.
1. The experiment consists of a series of Bernoulli trials.
2. The trials are identical and independent. The probability of success,
p, remains the same for each trial.
3. The trials are observed until exactly r successes are obtained, where r
is fixed beforehand.
4. The random variable X is the number of trials needed to obtain the r
successes.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 83 / 355
Elements of Probability Theory Discrete Random Variables
We will try to derive a formula for the probability density associated with
negative binomial trials. First, notice that if we need x trials for r
successes, then in x − 1 trials we must have had exactly r − 1 successes
and therefore x − r failures. We know that the probability of this is
µ ¶
x − 1 r −1
P[exactly r − 1 successes in x − 1 trials] = p (1 − p)x−r .
r −1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 84 / 355
Elements of Probability Theory Discrete Random Variables
Proof.
We derive the moment-generating function only. It is given by
∞
X ¶ µ
x −1 r
mX (t) = E[e ] = Xt
e p (1 − p)x−r
tx
x=r
r − 1
X∞ µ ¶
t(r +x) r + x − 1
= e p r (1 − p)x
r −1
x=0
X∞ µ ¶
r −1+x
r tr
=p e [e t (1 − p)]x
x
x=0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 86 / 355
Elements of Probability Theory Discrete Random Variables
where
µ ¶
−r (−r ) · (−r − 1) · · · (−r − x + 1)
=
x x!
r · (r + 1) · · · (r + x − 1)
= (−1)x
x! µ ¶
x (r + x − 1)! x r −1+x
= (−1) = (−1)
x!(r − 1)! x
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 87 / 355
Elements of Probability Theory Discrete Random Variables
Proof (continued).
∞ µ
X ¶
−r r −1+x x
It follows that (1 − y ) = y . Therefore,
x
x=0
∞ µ
X ¶
r −1+x
mX (t) = p e r tr
[e t (1 − p)]x
x
x=0
(pe t )r
= p r e tr (1 − (1 − p)e t )−r =
(1 − qe t )r
with q = 1 − p.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 88 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 89 / 355
Elements of Probability Theory Discrete Random Variables
1.2.25. Remark. If the sampling were done with replacement, this would
correspond to a binomial situation.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 90 / 355
Elements of Probability Theory Discrete Random Variables
In order to find a density function for this situation, we use the classical
definition of probability. Fix the number of objects N, the sample size n
and the number of objects with the trait, r .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 91 / 355
Elements of Probability Theory Discrete Random Variables
X : S → Ω = {x ∈ N : max(0, n − (N − r )) ≤ x ≤ min(n, r )}
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 92 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 93 / 355
Elements of Probability Theory Discrete Random Variables
N = # of items in a lot,
and
r = # of defectives in a lot.
Then we can calculate the probability that a sample of size n contains x
defectives.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 94 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 95 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 96 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 97 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 98 / 355
Elements of Probability Theory Discrete Random Variables
We also assume that there exists some number λ > 0 such that for any
small time interval of size ∆t the following postulates are satisfied:
1. The probability that exactly one arrival will occur in an interval of
width ∆t is approximately λ · ∆t.
2. The probability that exactly zero arrivals will occur in the interval is
approximately 1 − λ · ∆t.
3. The probability that two or more arrivals ocur in the interval is
approximately zero (very small).
We will formulate these postulates mathematically, and from them obtain
the density function for the Poisson distribution.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 99 / 355
Elements of Probability Theory Discrete Random Variables
f (t)
lim = 0.
t→0 t
Hence o(t) does not denote a particular function, rather a class of
functions. For example,
◮ t 2 = o(t),
◮ (1 + t)2 = 1 + 2t + o(t),
◮ sin t = t + o(t).
In particular,
◮ o(t) + o(t) = o(t),
◮ t n · o(t) = o(t) for all n ∈ N,
◮ o(t) · o(t) = o(t).
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 100 / 355
Elements of Probability Theory Discrete Random Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 101 / 355
Elements of Probability Theory Discrete Random Variables
so that
p0 (t + ∆t) − p0 (t) o(∆t)
−λp0 (t) = + .
∆t ∆t
We can take the limit as ∆t → 0 on both sides. Then the fraction
o(∆t)/∆t vanishes by definition, and we have
p0 (t + ∆t) − p0 (t)
−λp0 (t) = lim = p0′ (t).
∆t→0 ∆t
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 102 / 355
Elements of Probability Theory Discrete Random Variables
so that
px (t + ∆t) − px (t) o(∆t)
λpx−1 (t) − λpx (t) = + .
∆t ∆t
We can take the limit as ∆t → 0 on both sides. Then the fraction
o(∆t)/∆t vanishes by definition, and we have
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 103 / 355
Elements of Probability Theory Discrete Random Variables
(λt)x −λt
fXt (x) = px (t) = e .
x!
If we set k = λt, we obtain the Poisson distribution with parameter k.
1.2.30. Definition. Let k ∈ R. A random variable (X , fX ) with
X: S →N
k x e −k
fX (x) =
x!
is said to have a Poisson distribution with parameter k.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 104 / 355
Elements of Probability Theory Discrete Random Variables
2. E[X ] = k.
3. Var X = k.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 105 / 355
Elements of Probability Theory Discrete Random Variables
k = pn.
Note that usually k = λt, where λ corresponds to the arrivals per unit
time. In general, we require p < 0.1 for this approximation. The smaller p
and the larger n are, the better the approximation.
In fact, the Poisson distribution can be regarded as a limiting case of the
binomial distribution, as you shall prove in the exercises.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 107 / 355
Elements of Probability Theory Discrete Random Variables
1.2.33. Example. The probability that a particular rivet in the wing surface
of a new aircraft is defective is 0.001. There are 4000 rivets in the wing.
What is the probability that not more than six defective rivets will be
installed?
The true probability is
6 µ
X ¶
4000
P[X ≤ 6] = (0.001)x (0.999)4000−x .
x
x=0
X
6
4k
P[X ≤ 6] ≈ e −4 = 0.889.
x!
x=0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 108 / 355
Elements of Probability Theory Continuous Random variables
Joint Distributions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 109 / 355
Elements of Probability Theory Continuous Random variables
The function fX is called the probability density function (or just density)
of the random variable X .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 110 / 355
Elements of Probability Theory Continuous Random variables
Cumulative Distribution
Notice that by the above definition,
Z x
P[X = x] = fX (y ) dy = 0,
x
(X , fX ) = (Y , fY ) almost surely.
Cumulative Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 112 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 113 / 355
Elements of Probability Theory Continuous Random variables
Exponential Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 114 / 355
Elements of Probability Theory Continuous Random variables
Exponential Distribution
2 Β=2
3
Β=
2
Β=1
1
1 Β=
2
1
2
x
-1 1 2 3
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 115 / 355
Elements of Probability Theory Continuous Random variables
Exponential Distribution
We can easily calculate the expectation and variance for the exponential
distribution:
Z ∞ Z ∞
x −x/β
E[X ] = x fβ (x) dx = e dx
−∞ 0 β
Z ∞
¯
−x/β ¯∞
= −xe 0
+ e −x/β dx = β
0
Z ∞ Z ∞
2 2 x 2 −x/β
E[X ] = x fβ (x) dx = e dx
−∞ 0 β
Z
¯∞ ∞
= −2xe −x/β ¯0 + 2 x e −x/β dx = 2β 2
0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 116 / 355
Elements of Probability Theory Continuous Random variables
Exponential Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 117 / 355
Elements of Probability Theory Continuous Random variables
Exponential Distribution
The exponential distribution has an interpretation as “time-to-failure.” In
fact, there is a relationship between the exponential distribution and the
Poisson distribution.
Recall that for Poisson-distributed events (arrivals) the probability of x
arrivals in the time interval [0, t] was given by
(λt)x −λt
p(x) = e , x ∈ N.
x!
Then p(0) is the probability of no arrivals in [0, t]. This can also be
interpreted as the probability that the first arrival occurs at a time greater
than t. Let the time of the first arrival be a continuous random variable
denoted by T . Then
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 118 / 355
Elements of Probability Theory Continuous Random variables
Exponential Distribution
fT (t) = λe −λt , t ≥ 0.
and fT (t) = 0 for t < 0. Thus the time between successive arrivals of a
Poisson-distributed random variable is exponentially distributed with
parameter β = 1/λ.
If we interpret an “arrival” as the failure of a component, then β −1 = λ
represents the rate of failure.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 119 / 355
Elements of Probability Theory Continuous Random variables
Exponential Distribution
That is, 63.2% of the components will fail before the mean life time. As
you can see, this result does not depend on the value of β.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 120 / 355
Elements of Probability Theory Continuous Random variables
Exponential Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 121 / 355
Elements of Probability Theory Continuous Random variables
Exponential Distribution
Then
P[(X > x + s) ∩ (X > x)] P[X > x + s]
P[X > x + s | X > x] = =
P[X > x] P[X > x]
e −λ(x+s)
= = e −λs = P[X > s].
e −λx
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 122 / 355
Elements of Probability Theory Continuous Random variables
Gamma Distribution
A more general form of the exponential distribution is the gamma
distribution, which has two parameters.
1.3.6. Definition. Let α, β ∈ R, α, β > 0. A continuous random variable
(X , fα,β ) with density
(
1 α−1 e −x/β , x > 0,
αx
fα,β (x) = Γ(α)β
0, x ≤ 0,
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 123 / 355
Elements of Probability Theory Continuous Random variables
Gamma Distribution
The gamma function satisfies Γ(1) = 1 and Γ(α) = (α − 1)Γ(α − 1) if
α > 1. In other words,
n! = Γ(n + 1) for n ∈ N.
It is a continuous extension of the factorial function to the positive real
numbers and can in fact be defined for all complex numbers except zero
and the negative integers. Below is its graph for α ∈ (−3, 5).
y
15
y = G HxL = à
¥
10 z x - 1 e-z âz
0
x
-3 -2 -1 1 2 3 4 5
-5
-10
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 124 / 355
Elements of Probability Theory Continuous Random variables
Gamma Distribution
2. E[X ] = αβ.
3. Var X = αβ 2 .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 125 / 355
Elements of Probability Theory Continuous Random variables
Gamma Distribution
Proof.
We will verify the moment-generating function only.
Z ∞
e tx
mX (t) = E[e tX ] = α
x α−1 e −x/β dx
Γ(α)β
Z 0∞
1
= x α−1 e −x(1/β−t) dx
Γ(α)β α 0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 126 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 127 / 355
Elements of Probability Theory Continuous Random variables
0.15
0.10
0.05
5 10 15 20
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 128 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 129 / 355
Elements of Probability Theory Continuous Random variables
Normal Distribution
- J N
y = H2 ΠL-12 Σ-1 ã
1 x-Μ 2
2 Σ
Μ-Σ Μ Μ+Σ
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 130 / 355
Elements of Probability Theory Continuous Random variables
Normal Distribution
R
It is easily verified that R fX (x) dx = 1 by using polar coordinates (see
blackboard). We furthermore have the following result:
1.3.10. Theorem. Let (X , fX ) be a normally distributed random variable
with parameters µ and σ.
1. The moment-generating function of X is given by
2 t 2 /2
mX : R → R, mX (t) = e µt+σ .
2. E[X ] = µ.
3. Var X = σ 2 .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 131 / 355
Elements of Probability Theory Continuous Random variables
Normal Distribution
Proof.
We will verify the moment-generating function only.
Z ∞
e tx −((x−µ)/σ)2
tX
mX (t) = E[e ] = √ e dx
−∞ 2πσ
Z ∞
1 2
=√ e tx−((x−µ)/σ) /2 dx
2πσ −∞
We complete the square in the exponent to gain
(x − (µ + σ 2 t))2
tx − ((x − µ)/σ)2 /2 = − + µt + σ 2 t 2 /2
2σ 2
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 132 / 355
Elements of Probability Theory Continuous Random variables
Normal Distribution
Proof (continued).
Substituting into the integral,
Z ∞
1 (x−(µ+σ 2 t))2
mX (t) = √ e− 2σ 2
+µt+σ 2 t 2 /2
dx
2πσ −∞
Z ∞
1 (x−(µ+σ 2 t))2
=e µt+σ 2 t 2 /2
√ e− 2σ 2 dx
2πσ −∞
| {z }
=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 133 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 134 / 355
Elements of Probability Theory Continuous Random variables
It is easily seen that Z = X σ−µ has mean E[Z ] = 0 and variance Var Z = 1,
but it is not clear that Z is normally distributed. To see this, we need to
find the density of Z .
Hence it is worth studying the density of transformed variables in general.
1.3.13. Theorem. Let X be a continuous random variable with density fX .
Let Y = φ ◦ X , where φ : R → R is strictly monotonic and differentiable.
The density for Y is then given by
¯ −1 ¯
¯ dφ (y ) ¯
−1 ¯
fY (y ) = fX (φ (y )) · ¯ ¯.
dy ¯
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 135 / 355
Elements of Probability Theory Continuous Random variables
FY (y ) = P[Y ≤ y ] = P[φ(X ) ≤ y ].
dφ−1 (z)
φ−1 (z) = σz + µ, = σ > 0.
dz
Using
1
e −((x−µ)/σ) /2
2
fX (x) = √
2πσ
we have
¯ −1 ¯
¯ dφ (z) ¯
fZ (z) = fX (φ −1
(z)) · ¯¯ ¯ = √ 1 e −(z)2 /2 · σ = √1 e −z 2 /2 ,
dz ¯ 2πσ 2π
which is the density of the standard normal distribution. Hence the
variable Z = x−µ
σ is standard normal.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 137 / 355
Elements of Probability Theory Continuous Random variables
0.5
0.1
0.4
0.08
0.06 0.3
0.02 0.1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 139 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 140 / 355
Elements of Probability Theory Continuous Random variables
With the adjustment, 98.76% of the product will fall inside tolerance.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 141 / 355
Elements of Probability Theory Continuous Random variables
Σ=0.0004
Σ=0.0004
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 144 / 355
Elements of Probability Theory Continuous Random variables
Hence 68% of the values of a normal random variable lie within one
standard deviation of the mean, 95% lie within two standard deviations,
and 99.7% lie within three standard deviations. This rule of thumb will be
especially important in statistics, where the number of “extraordinary”
events needs to be judged.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 145 / 355
Elements of Probability Theory Continuous Random variables
Chebyshev’s Inequality
There is a general estimate, which does not depend on the distribution of
a discrete or continuous random variable, that tells us how the variance
influences the probability of deviating from the mean. This estimate is
called Chebyshev’s inequality.
1.3.19. Theorem. Let (X , fX ) be a discrete or continuous random variable
and k > 0 some positive number. Then
1
P[−kσ < X − µ < kσ] ≥ 1 − (1.3.1)
k2
or, equivalently,
1
P[|X − µ| ≥ kσ] ≤ . (1.3.2)
k2
Comparing (1.3.2) with Theorem 1.3.18, we see that the estimates in the
theorem are better. This is not surprising, as Chebyshev’s rule is valid for
any distribution, while the previous theorem uses the specific properties of
the normal distribution.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 146 / 355
Elements of Probability Theory Continuous Random variables
Chebyshev’s Inequality
Proof.
We will prove the theorem for the case of a continuous random variable
only. The proof in the discrete case is quite similar. By definition,
Z
σ = Var X = E[(X − µ) ] = (x − µ)2 fX (x) dx
2 2
R
Z √ Z √
µ− K µ+ K
= (x − µ)2 fX (x) dx + √ (x − µ)2 fX (x) dx
−∞ µ− K
| {z }
≥0
Z ∞
+ √ (x − µ)2 fX (x) dx
µ+ K
Chebyshev’s Inequality
Proof (continued).
√
Now (x −√µ)2 ≥ K if and√only if |x − µ| ≥ K which is the case if
x ≥ µ + K or x ≤ µ − K . Therefore,
Z √ Z
µ− K ∞
σ ≥
2
(x − µ) fX (x) dx +
2
√ (x − µ)2 fX (x) dx
−∞ µ+ K
Z √ Z
µ− K ∞
≥K fX (x) dx + K √ fX (x) dx
−∞ µ+ K
³ √ √ ´
= K P[X ≤ µ − K ] + P[X ≥ µ + K ] ,
or, equivalently,
√ σ2
P[|X − µ| ≥ K] ≤ .
K
√ 1
Taking k = K /σ, we obtain P[|X − µ| ≥ kσ] ≤ .
k2
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 148 / 355
Elements of Probability Theory Continuous Random variables
Chebyshev’s Inequality
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 149 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 150 / 355
Elements of Probability Theory Continuous Random variables
in the sense that the relative error converges to zero as n goes to infinity,
√
n! − 2πe −n nn+1/2
lim = 0.
n→∞ n!
Then for a binomial random variable with parameters n and p (q = 1 − p)
we can eventually show that
n! 1
√ e −(x−np) /(2npq)
2
P[X = x] = p x q n−x ≈ √
x!(n − x)! npq 2π
so that µ ¶
x − np
P[X ≤ x] ≈ Φ √
npq
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 151 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 152 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 153 / 355
Elements of Probability Theory Continuous Random variables
Weibull Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 154 / 355
Elements of Probability Theory Continuous Random variables
Weibull Distribution
1.3.22. Definition. A random variable (X , fX ) is said to have a
(two-parameter) Weibull distribution with parameters α and β if its
density is given by
(
αβx β−1 e −αx , x > 0,
β
f (x) = α, β > 0.
0, otherwise,
For the proof of the expression for the mean, we refer to the text book.
The proof of the formula for the variance is left as an exercise.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 155 / 355
Elements of Probability Theory Continuous Random variables
Weibull Distribution
The Weibull distribution is used (for example)
◮ In reliability engineering and failure analysis (the most common usage)
◮ In survival analysis (a branch of statistics which deals with death in
biological organisms and failure in mechanical systems)
◮ To represent manufacturing and delivery times in industrial
engineering
◮ In weather forecasting
◮ In radar systems to model the dispersion of the received signals level
produced by some types of clutters
◮ To model fading channels in wireless communications
◮ In general insurance to model the size of reinsurance claims
◮ In forecasting technological change (also known as the Sharif-Islam
model)
◮ To describe wind speed distributions, as the natural distribution often
matches the Weibull shape
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 156 / 355
Elements of Probability Theory Continuous Random variables
Reliability
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 157 / 355
Elements of Probability Theory Continuous Random variables
Reliability
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 158 / 355
Elements of Probability Theory Continuous Random variables
Note that
P[X ∈ [t, t + ∆t]] F (t + ∆t) − F (t)
lim = lim = F ′ (t) = f (t)
∆t→0 ∆t ∆t→0 ∆t
and P[X ∈ [t, ∞]] = 1 − F (t) = R(t) does not depend on ∆t. Therefore,
f (t)
ϱ(t) = .
R(t)
The job of the scientist is to find the form of these functions for the
problem at hand. In practice, one often begins by assuming a particular
form for the hazard rate function based on empirical evidence.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 159 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 160 / 355
Elements of Probability Theory Continuous Random variables
Reliability
Often one has an idea of ϱ, but not of the failure density f or reliability
function R. Then the following theorem enables one to determine the
other two functions:
1.3.24. Theorem. Let X be a random variable with failure (probability)
density X , reliability function R and hazard rate ϱ. Then
Rt
R(t) = e − 0 ϱ(x) dx
Proof.
Note that since R(x) = 1 − F (x) we have R ′ (x) = −F ′ (x). Therefore,
so R ′ (x) = −ϱ(x)R(x). Note that R(0) = 1, since the component will not
fail before t > 0. Solving this differential equation through separation of
variables, we obtain the result.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 161 / 355
Elements of Probability Theory Continuous Random variables
Reliability
1.3.25. Example. One hazard function in widespread use is the function
ϱ(t) = αβt β−1 , t > 0, α, β > 0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 163 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 165 / 355
Elements of Probability Theory Continuous Random variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 166 / 355
Elements of Probability Theory Joint Distributions
Joint Distributions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 167 / 355
Elements of Probability Theory Joint Distributions
Random Vectors
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 168 / 355
Elements of Probability Theory Joint Distributions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 169 / 355
Elements of Probability Theory Joint Distributions
X = x/Y = y 0 1 2 3
0 0.840 0.030 0.020 0.010
1 0.060 0.010 0.008 0.002
2 0.010 0.005 0.004 0.001
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 170 / 355
Elements of Probability Theory Joint Distributions
1.4.5. Example.
X = x/Y = y 0 1 2 3 fX (x)
0 0.840 0.030 0.020 0.010 0.900
1 0.060 0.010 0.008 0.002 0.080
2 0.010 0.005 0.004 0.001 0.020
fY (y ) 0.910 0.045 0.032 0.013 1.00
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 172 / 355
Elements of Probability Theory Joint Distributions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 173 / 355
Elements of Probability Theory Joint Distributions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 174 / 355
Elements of Probability Theory Joint Distributions
Independence
Two events A1 , A2 are independent if P[A1 ∩ A2 ] = P[A1 ] · P[A2 ]. This
motivates us to define the independence of two discrete random variables
through
P[X = x and Y = y ] = P[X = x] · P[Y = y ]
which works out to fXY (x, y ) = fX (x)fY (y ). This also generalizes to
continuous random variables.
1.4.8. Definition. Let ((X , Y ), fXY ) be a bivariate random variable with
marginal densities fX and fY . If
and
Independence
1.4.9. Example.
X = x/Y = y 0 1 2 3 fX (x)
0 0.840 0.030 0.020 0.010 0.900
1 0.060 0.010 0.008 0.002 0.080
2 0.010 0.005 0.004 0.001 0.020
fY (y ) 0.910 0.045 0.032 0.013 1.00
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 176 / 355
Elements of Probability Theory Joint Distributions
Conditional Densities
Whenever we consider two random variables, it does not make sense to
speak of (X , fX ) and (Y , fY ) anymore, as the two random variables may
influence each other. We therefore always need to consider a joint density
fXY for the pair (X , Y ), i.e., a bivariate random variable. The “individual
densities” fX and fY are then just the marginal densities.
If Y = y is kept fixed, the density of the random variable (X , fX ) will be
some given function for this value of Y . It may be another function for
another value of Y .
The conditional probability for an event A1 given A2 was
P[A1 | A2 ] = P[A1 ∩ A2 ]/P[A2 ]. This motivates us to define the
independence of two discrete random variables through
Conditional Densities
fXY (x, y )
fX |y (x) =
fY (y )
whenever fY (y ) > 0.
2. The conditional density for Y given X = x is defined as
fXY (x, y )
fY |x (y ) =
fX (x)
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 178 / 355
Elements of Probability Theory Joint Distributions
Note that this implies E[X + Y ] = E[X ] + E[Y ], as we have stated earlier.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 179 / 355
Elements of Probability Theory Joint Distributions
X = x/Y = y 0 1 2 3 fX (x)
0 0.840 0.030 0.020 0.010 0.900
1 0.060 0.010 0.008 0.002 0.080
2 0.010 0.005 0.004 0.001 0.020
fY (y ) 0.910 0.045 0.032 0.013 1.00
X X
2 X
3
E[X ] = x · fXY (x, y ) = x · fXY (x, y ) = 0.12
(x,y )∈Ω x=0 y =0
X X
2 X
3
E[Y ] = y · fXY (x, y ) = y · fXY (x, y ) = 0.148
(x,y )∈Ω x=0 y =0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 180 / 355
Elements of Probability Theory Joint Distributions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 181 / 355
Elements of Probability Theory Joint Distributions
Covariance
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 182 / 355
Elements of Probability Theory Joint Distributions
Covariance
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 183 / 355
Elements of Probability Theory Joint Distributions
Correlation
While the covariance tells us whether or not two random variables affect
each other, there is no quantitative statement contained in the covariance
as such. We often want to know, however, if there is a linear dependence
between X and Y . This can be determined from the Pearson coefficient of
correlation, defined as follows.
Cov(X , Y ) σXY
ρXY = p = .
(Var X )(Var Y ) σ X σY
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 184 / 355
Elements of Probability Theory Joint Distributions
Correlation
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 185 / 355
Elements of Probability Theory Joint Distributions
Correlation
Proof.
We first show that −1 ≤ ρXY ≤ 1.
We consider two random variables W and Z such that E[Z 2 ], E[W 2 ] ̸= 0.
Note that for any a ∈ R (aW − Z )2 ≥ 0, so
E[WZ ]2 E[WZ ]2
− + E[Z 2 ] ≥ 0 ⇔ ≤ 1.
E[W 2 ] E[W 2 ] E[Z 2 ]
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 186 / 355
Elements of Probability Theory Joint Distributions
Correlation
Proof (continued).
Now let ρ2XY = 1. Then, reversing the steps above, we obtain
Y = (µY − aµX ) + aX
almost surely.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 187 / 355
Elements of Probability Theory Joint Distributions
Transformation of Variables
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 188 / 355
Elements of Probability Theory Joint Distributions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 189 / 355
An Introduction to Statistical Methods
Part II
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 190 / 355
An Introduction to Statistical Methods
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 191 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 192 / 355
An Introduction to Statistical Methods Descriptive statistics
Descriptive Statistics
Up to now we have discussed probability theory, i.e., the properties of
random variables and their distributions. It is clear that we have only
scratched the surface of probability theory and a lot more can be explored.
For example, the theory of stochastic processes and Markov chains have
not been discussed.
Now, however, instead of delving deeper into probability theory, we will
leave the “perfect world” of known random variables and distributions,
and enter the “real world” of statistics, which deals with incomplete
information. Statistical problems are characterized through
◮ a large group of objects about which inferences are to be made, called
a population,
◮ at least one random variable whose behavior is to be studied relative
to the population,
◮ a subset of the population, called a sample which is actually studied.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 193 / 355
An Introduction to Statistical Methods Descriptive statistics
2.1.1. Examples.
1. We may want to find the average gas mileage of cars in Shanghai.
Then the population is “all cars in Shanghai”, and we may pick a
sample of 100 cars. Here the random variable is the gas mileage, of
which we wish to obtain the mean.
2. We may want to find out whether a proposed new car model will have
a lower mean gas mileage than existing cars. In this case, the
population is “all cars of this model, existing now and produced in the
future” and a sample might consist of a trial production of 20
prototype cars. Again the random variable is the gas mileage and we
are interested in its expected value.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 194 / 355
An Introduction to Statistical Methods Descriptive statistics
Random Sampling
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 195 / 355
An Introduction to Statistical Methods Descriptive statistics
Random Sampling
However, we define:
2.1.2. Definition. A random sample of size n from the distribution of X is
a collection of n independent random variables X1 , . . . , Xn , each with the
same distribution as X . We say they are independent identically
distributed (i.i.d.) random variables.
In order to guarantee that the the random variables in a random sample
are indeed independently distributed, the size of the random sample should
not exceed 5% of the population.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 196 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 197 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 198 / 355
An Introduction to Statistical Methods Descriptive statistics
Stem-and-Leaf Diagrams
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 199 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 200 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 201 / 355
An Introduction to Statistical Methods Descriptive statistics
Sample Range
max Xk − min Xk
1≤k≤n 1≤k≤n
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 202 / 355
An Introduction to Statistical Methods Descriptive statistics
Histograms
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 203 / 355
An Introduction to Statistical Methods Descriptive statistics
Histograms
Steps in creating a histogram for numerical data:
1. Find the desired number of categories:
Data Set Size # Categories
< 16 Insufficient data
2 n−1 to 2n − 1 n
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 204 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 205 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 206 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 207 / 355
An Introduction to Statistical Methods Descriptive statistics
Ogives
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 208 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 209 / 355
An Introduction to Statistical Methods Descriptive statistics
1X
n
X = Xi
n
i=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 210 / 355
An Introduction to Statistical Methods Descriptive statistics
P[X ≤ M] = 0.50.
This sample median may differ significantly from the sample mean!
We define the median location by the number (n + 1)/2.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 211 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 212 / 355
An Introduction to Statistical Methods Descriptive statistics
1X
n
S2 = (Xk − X )2 .
n
k=1
1 X
n
S2 = (Xk − X )2 ,
n−1
k=1
defining
√ the sample variance in this way and the sample standard deviation
S = S 2.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 213 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 214 / 355
An Introduction to Statistical Methods Descriptive statistics
Rounding of Statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 215 / 355
An Introduction to Statistical Methods Descriptive statistics
Boxplots - Quartiles
Boxplots are a very useful way of visualizing data. In order to construct a
boxplot, we first need to determine the quartiles q1 and q3 and the
interquartile range iqr = q3 − q1 . The quartiles play a similar role to that
of the median (which would be the quartile denoted q2 ) in that of a
random sample ordered from smallest to largest, 25% would lie below q1
and 75% would lie below q3 . The precise construction of q1 and q3 varies
(one algorithm is given in the textbook).
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 216 / 355
An Introduction to Statistical Methods Descriptive statistics
Construction of Boxplots
The construction of a boxplot will be demonstrated on the blackboard.
We need the following data:
◮ q1 , e
x , q3 and iqr.
◮ Inner fences
3 3
f 1 = q1 − iqr, f3 = q3 + iqr .
2 2
◮ Adjacent values
a1 = min{xk : xk ≥ f1 }, a3 = max{xk : xk ≤ f3 }.
◮ Outer fences
F1 = q1 − 3 iqr, F3 = q3 + 3 iqr .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 217 / 355
An Introduction to Statistical Methods Descriptive statistics
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 218 / 355
An Introduction to Statistical Methods Descriptive statistics
Outliers
Data points lying between the inner and outer fences are called near
outliers, those lying outside the outer fences are called far outliers. Far
outliers are unusual if (and only if!) an approximately bell-shaped
distribution of the random variable X of the population is expected. In
this case, their origin should be investigated.
◮ If the outlier seems to be the result of an error in measurement or
data collecting, it may be discarded from the data.
◮ If the outlier seems to be the result of a random measurement, it is
recommended that statistics are reported twice: with the outlier
included and without the outlier.
◮ As a rule of thumb: Of 1000 random samples of a normally
distributed population, it can be expected that 7 will be outliers.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 219 / 355
An Introduction to Statistical Methods Estimation
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 220 / 355
An Introduction to Statistical Methods Estimation
Estimation
In the last section we have seen how data obtained from a random sample
can be used to obtain information on a population; in particular a statistic
(such as the median) would approximate a random variable (such as the
mean). However, no precise information on the “quality” of the
approximation was given, and one formula (for the sample variance)
remained obscure and counterintuitive.
The process of using statistics to approximate random variables is called
estimation. We now aim to provide a mathematical framework for this
process. Note that the language of statistics differs slightly from that of
probability theory; instead of a random variable X of a population, we
refer to a more general population parameter θ such as the mean or
standard deviation. But a population parameter can also be the parameter
of a certain distribution (such as λ of the Poisson distribution). We have
previously seen that functions of random samples X1 , . . . , Xn (which in
probability theory are random variables themselves) are called statistics.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 221 / 355
An Introduction to Statistical Methods Estimation
Estimators
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 222 / 355
An Introduction to Statistical Methods Estimation
Estimators
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 223 / 355
An Introduction to Statistical Methods Estimation
Sample Mean
Proof.
We simply insert the definition of the sample mean and use the properties
of the expectation:
1
E[X ] = E[(X1 + · · · + Xn )/n] = E[X1 + · · · + Xn ]
n
1 nµ
= (E[X1 ] + · · · + E[Xn ]) = = µ.
n n
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 224 / 355
An Introduction to Statistical Methods Estimation
Sample Variance
2.2.3. Theorem. Let X be the sample mean of a random sample of size n
from a distribution with mean µ and variance σ 2 . Then
σ2
Var X = E[(X − µ)2 ] = .
n
Proof.
We simply insert the definition of the sample mean and use the properties
of the variance:
1
Var X = Var((X1 + · · · + Xn )/n) = Var(X1 + · · · + Xn )
n2
1 nσ 2 σ2
= 2 (Var X1 + · · · + Var Xn ) = 2 = .
n n n
Thus X is both unbiased and has a variance that decreases with large n; it
is a “nice” estimator, since we can make the mean square error MSE X as
small as desired by taking n large enough.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 225 / 355
An Introduction to Statistical Methods Estimation
√ √
2.2.4. Definition. The standard deviation of X is given by Var X = σ/ n
and is called the standard error of the mean.
1 X
n
S =2
(Xk − X )2 ,
n−1
k=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 226 / 355
An Introduction to Statistical Methods Estimation
Sample Variance
Proof.
We simply calculate E[S 2 ],
· ¸
1 X
n
2
E[S ] = E (Xk − X) 2
n−1
k=1
·Xn ¸
1
= E (Xk − µ + µ − X) 2
n−1
k=1
·Xn X
n X
n ¸
1
= E (Xk − µ)2 − 2(X − µ) (Xk − µ) + (µ − X )2
n−1
k=1 k=1 k=1
·Xn ³³X
n ´ ´ ¸
1
= E (Xk − µ) − 2(X − µ)
2
Xk − nµ + n(µ − X ) .
2
n−1
k=1 k=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 227 / 355
An Introduction to Statistical Methods Estimation
Sample Variance
Proof (continued).
Pn
Note that k=1 Xk = nX , so
·X n ³³X n ´ ´ ¸
1
2
E[S ] = E (Xk − µ) − 2(X − µ)
2
Xk − nµ + n(µ − X )2
n−1
k=1 k=1
·X n ¸
1
= E (Xk − µ)2 − 2(X − µ)(nX − nµ) + n(µ − X )2
n−1
k=1
·X n ¸
1
= E (Xk − µ) − n(X − µ)
2 2
n−1
k=1
µX n ¶
1
= E[(Xk − µ) ] − n E[(X − µ) ] .
2 2
n−1
k=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 228 / 355
An Introduction to Statistical Methods Estimation
Sample Variance
Proof (continued).
Since Var Xk = σ 2 = E[(Xk − µ)2 ] for each k = 1, . . . , n, and
E[(X − µ)2 ] = σ 2 /n by Theorem 2.2.3, we have
µXn ¶
1
2
E[S ] = E[(Xk − µ) ] − n E[(X − µ) ]
2 2
n−1
k=1
µXn ¶
1 σ2
= σ −n
2
n−1 n
k=1
1
= (nσ 2 − σ 2 ) = σ 2 .
n−1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 229 / 355
An Introduction to Statistical Methods Estimation
The method of moments goes back to Karl Pearson in 1894 and uses the
basic fact that unbiased estimators Mk for the kth moments E [X k ] of a
distribution are
1X k
n
Mk = Xi ,
n
i=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 230 / 355
An Introduction to Statistical Methods Estimation
M2 − M12 M1 M12
β̂ = , α̂ = = .
M1 β̂ M2 − M12
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 231 / 355
An Introduction to Statistical Methods Estimation
2.2.7. Example. Water samples of a specific size are taken from a river
suspected of having been polluted by improper treatment procedures at an
upstream sewage-disposal plant. Let X denote the number of coliform
organism found per sample, and assume that X is a Poisson random
variable with parameter k. Let x1 , . . . , xn be a random sample from the
distribution of X . We want to determine the value of k that gives the
highest probability of observing this sample.
Since random sampling implies independence,
Y
n
P[X1 = x1 and X2 = x2 and . . . and Xn = xn ] = P[Xj = xj ].
j=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 233 / 355
An Introduction to Statistical Methods Estimation
Maximizing ln L(k) will also maximize L(k), so we take the first derivative
and set it equal to zero:
1X
n
d ln L(k)
= −n + xj = 0 ⇔ k = x.
dk k
j=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 234 / 355
An Introduction to Statistical Methods Estimation
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 235 / 355
An Introduction to Statistical Methods Estimation
mY (t) = e αt mX (βt).
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 236 / 355
An Introduction to Statistical Methods Estimation
Confidence Intervals
2.2.11. Definition. Let 0 ≤ α ≤ 1. A 100(1 − α)% confidence interval for a
parameter θ is an interval [L1 , L2 ] such that
P[L1 ≤ θ ≤ L2 ] = 1 − α.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 237 / 355
An Introduction to Statistical Methods Estimation
Interval Estimation
y =x ±ε :⇔ y ∈ [x − ε, x + ε].
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 238 / 355
An Introduction to Statistical Methods Estimation
Interval Estimation
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 239 / 355
An Introduction to Statistical Methods Estimation
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 240 / 355
An Introduction to Statistical Methods Estimation
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 241 / 355
An Introduction to Statistical Methods Estimation
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 242 / 355
An Introduction to Statistical Methods Estimation
Interval Estimation
2.2.14. Example. An article in the Journal of Heat Transfer describes a
method of measuring the thermal conductivity of Armco iron. Using a
temperature of 100◦ F and a power input of 550 W, the following 10
measurements of thermal conductivity (in Btu /(hr ft ◦ F)) were obtained:
41.60 41.48 42.34 41.95 41.86
42.18 41.72 42.26 41.81 42.04
A point estimate of the mean thermal conductivity at 100◦ F and 550 W is
the sample mean,
x = 41.92 Btu /(hr ft ◦ F).
Suppose we know that the standard deviation of the thermal conductivity
under the given conditions is σ = 0.10 Btu /(hr ft ◦ F). A 95% confidence
interval (α = 0.05) on the mean is then given by
z0.025 · σ
x± √ = 41.924 ± 0.062 = [41.862, 41.986].
n
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 243 / 355
An Introduction to Statistical Methods Estimation
While up to now most assertions we have made have been proven (more or
less rigorously), many important results in statistics are highly non-trivial
and require an inordinate (for this course) amount of effort to prove. One
of the first of these is the Central Limit Theorem, which we now cite:
2.2.15. Theorem. Let X1 , . . . , Xn be a sequence of independent random
variables with arbitrary distributions, means E[Xj ] = µj and variances
Var Xj = σj2 (all finite). Let Y = X1 + · · · + Xn . Then under some general
conditions P
Y − µj
Zn = qP
σj2
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 244 / 355
An Introduction to Statistical Methods Estimation
Y − nµ X −µ
Zn = √ = √
σ n σ/ n
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 245 / 355
An Introduction to Statistical Methods Estimation
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 246 / 355
An Introduction to Statistical Methods Estimation
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 247 / 355
An Introduction to Statistical Methods Estimation
How large must n be for the central limit theorem to give a good
approximation?
This depends on how “well-behaved” the distributions of the variables Xj
are:
1. Well-behaved (nearly symmetric densities that look close to that of a
normal distribution): n ≥ 4.
2. Reasonably behaved (no prominent mode, densities look like uniform
densities): n ≥ 12.
3. Ill-behaved (much of the weight of the densities is in the tails,
irregular appearance): n ≥ 100.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 248 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 249 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Preliminary Theory
Fχn (y ) = P[χn ≤ y ].
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 250 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Preliminary Theory
For y > 0,
v
u n
1 u X
Fχn (y ) = P[χn ≤ y ] = P t (Xk − µ)2 < y
σ
k=1
" n µ ¶ # " n #
X Xk − µ 2 X
2 2 2
=P <y =P Zk < y
σ
k=1 k=1
Xk −µ
where the variables Zk = σ , k = 1, . . . , n, follow a standard normal
distribution.
The density for the sum of n independent normal distributions is the
product of their individual densities, so we have
Z Pn
(2π)−n/2 e − k=1 zk /2 dz1 . . . dzn .
2
Fχn (y ) =
Pn 2 2
k=1 zk <y
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 251 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Preliminary Theory
x1 = r sin θ1
x2 = r cos θ1 sin θ2
x3 = r cos θ1 cos θ2 sin θ3
..
.
xn−1 = r cos θ1 cos θ2 . . . cos θn−2 sin θn−1
xn = r cos θ1 cos θ2 . . . cos θn−2 cos θn−1
Here r > 0 and −π/2 < θk < π/2, k = 1, n − 2, 0 < θn−1 < 2π.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 252 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Preliminary Theory
The integral becomes
Z 2π Z π/2 Z π/2 Z y
(2π)−n/2 e −r
2 /2
Fχn (y ) = ... r n−1
0 −π/2 −π/2 0
× D(θ1 , . . . , θn−1 ) dr dθ1 . . . dθn−2 dθn−1
we have
Z y
e −r
2 /2
Fχn (y ) = Cn r n−1 dr .
0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 253 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Preliminary Theory
We determine Cn from
Z ∞ ³n´
e −r
2 /2
1 = lim Fχn (y ) = Cn r n−1 dr = Cn Γ 2n/2−1 .
y →∞ 0 2
Hence
Z y
1
e −r
2 /2
Fχn (y ) = ¡n¢ r n−1 dr .
Γ 2 2n/2−1 0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 254 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
The χ2 -Distribution
Next we consider the random variable
1 X
n
χ2n = (Xk − µ)2 . (2.3.1)
σ2
k=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 255 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Proof.
Each of the chi-squared random variables Xγ2k may be regarded as being a
sum of γk squares of standard-normally distributed random variables.
Pn
Hence their sum may be regarded as the sum of α = γk squares of
k=1
standard-normally distributed random variables, giving a chi-squared
random variable with α degrees of freedom.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 256 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
σ 2 = E[(X − µ)2 ]
is unknown, we must start all over again, and first learn more about the
sample variance
1 X
n
S2 = (Xk − X )2 .
n−1
k=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 258 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 259 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
or Y = AX for short. It is easy to see that the rows of the matrix A are
orthonormal. Thus, A is an orthogonal matrix, A−1 = AT . This
immediately implies |det A| = 1, since
1
det A = det AT = det A−1 = ⇒ (det A)2 = 1
det A
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 260 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
X
n X
n
yi2 = ⟨y, y⟩ = ⟨Ax, Ax⟩ = ⟨AT Ax, x⟩ = ⟨x, x⟩ = xi2 . (2.3.2)
i=1 i=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 261 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
√ Y
n
= (2π)−1/2 σ −1 e − 2σ2 (y1 − (2π)−1/2 σ −1 e − 2σ2 yi
1 1
nµ)2 2
i=2
Furthermore,
X
n X
n
2 X
n
(n − 1)S 2 = (Xi − X )2 = Xi2 − nX = Yi2 − Y12
i=1 i=1 i=1
X
n
= Yi2 .
i=2
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 263 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
it follows that √ √
−1/2 −1 −
( nx− nµ)2 √
fX (x) = (2π) σ e 2σ 2 n
so X is normally distributed with mean µ and variance σ 2 /n.
Now
1 X 2
n
(n − 1)S 2 /σ 2 = Yi
σ2
i=2
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 264 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
2.3.3. Remark. Theorem 2.3.2 essentially uses the fact that the i.i.d.
variable Xk , k = 1, . . . , n, are normally distributed. In fact, the converse
result is true also:
Let X1 , . . . , Xn , n ≥ 2, be i.i.d. random variables. Then if X and
S 2 are independent, the Xk , k = 1, . . . , n follow a normal
distribution.
This means that the independence of X and S 2 is a characteristic property
of the normal distribution. Furthermore, if in a given situation we assume
that X and S 2 are independently distributed we are essentially assuming
that the population is normally distributed.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 265 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
We can now use Theorem 2.3.2 to find a confidence interval for the
variance based on the sample variance S. First, for 0 < α ≤ 1 we define
χ21−α/2,n ≤ χ2α/2,n ∈ R by
Z χ21−α/2,n
fXn2 (x) dx = α/2,
0
Z ∞
fXn2 (x) dx = α/2,
χ2α/2,n
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 266 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 267 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 268 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 269 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
X −µ
Z= √
σ/ n
X −µ
√
S/ n
is not known.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 270 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 271 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
X −µ
Tn−1 = √
S/ n
Proof. √
We know that (X − µ)/(σ/ n) is standard normal and (n − 1)S 2 /σ 2 is a
chi-squared random variable with n − 1 degrees of freedom. Therefore,
√
Z (X − µ)/(σ/ n) X −µ
q = p = √
Xγ2 /γ ( (n − 1)S /σ )/(n − 1)
2 2 S/ n
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 273 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
x = 9.8475, s = 0.0954
Often, the statistician will have some idea of the value of a population
parameter, or will try to verify or refute a statement on this parameter. As
an example, a new type of battery might be designed to have a longer life
span than a traditional model. Then a series of prototypes of the new type
(constituting a sample from the population of all not yet produced
batteries of the new type) might be tested for their life span. If the mean
life span of the traditional batteries was 160 days, a hypothesis might be
that “the new type of batteries has a mean life span of more than 170
days”.
The hypothesis that is to be tested is called the research hypothesis and
denoted by H1 (in our example, H1 would be “µ > 170 days”), while the
negation of H1 is called the null hypothesis and denoted H0 (here:
“µ ≤ 170 days”.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 275 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Hypotheses
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 276 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Hypothesis Testing
We always test a pair of research and null hypotheses. Let θ denote a
population parameter whose value is being compared to some θ0 ∈ R.
Then we have
One-sided tests.
H 0 : θ ≤ θ0 , H1 : θ > θ 0
H 0 : θ ≥ θ0 , H1 : θ < θ 0
Two-sided test.
H 0 : θ = θ0 , H1 : θ ̸= θ0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 277 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Hypothesis Testing
On the basis of the test statistic, we either
◮ reject H0 or
◮ fail to reject H0 .
2.3.12. Example. We are given the task of testing the following hypothesis:
More than half of all car headlights in Shanghai are incorrectly
adjusted.
We now establish the mathematical/statistical framework for tackling this
problem:
◮ The population is the set of all cars in Shanghai;
◮ The random variable X is discrete: either a car has a correctly
adjusted headlight (X = 0) or it does not (X = 1);
◮ X follows a binomial distribution with
N = “number of cars in Shanghai” and population parameter p.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 278 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Hypothesis Testing
Our hypotheses are
H0 : p ≤ 0.5 H1 : p > 0.5
with the null value p0 = 0.5.
We take a random sample of n = 20 automobiles and count those that
have maladjusted headlights. Effectively, we are conducting an experiment
with a binomial random variable X : S → Ω = {1, . . . , 20}.
Our test statistic will be the number X of cars in the random sample with
incorrectly adjusted headlights.
If p = p0 , then X follows a binomial distribution with p = p0 = 0.5, and
we can expect p̂ = X /n to be close to p. In this case, E[X ] = np0 = 10.
We decide to reject H0 if at least 14 cars have incorrectly adjusted
headlights; the probability of this happening if p = p0 is
P[X ≥ 14 | p = 0.5 n = 20] = 1 − P[X ≤ 13 | p = 0.5 n = 20]
= 1 − 0.9423 = 0.0577.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 279 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Hypothesis Testing
Thus, if H0 is true (i.e., p = 0.5) then there is approximately a 6% chance
of observing 14 cars with incorrectly adjusted headlights.
What does this imply? Assume that we actually observe 14 or more cars
with incorrectly adjusted headlights and reject H0 . Then two there are two
possibilities:
◮ We have correctly rejected H0 or
◮ We have falsely rejected H0 .
In the second case, H0 is in fact true. However, the probability of this
happening is at most 6%, since if p ≤ p0 , then
Type I Errors
We want to keep the probability α of committing a Type I error as small
as possible. We directly control α by arbitrarily defining the critical region
of the test.
The critical region is the subset of the range of the test statistic which will
lead us to reject H0 . In our example, the critical region is
C = {14, 15, 16, 17, 18, 19, 20} ⊂ Ω.
If we reduce the critical region, we decrease α. For example, if we set
C = {16, 17, 18, 19, 20} ⊂ Ω,
i.e., we only reject H0 if at least 16 out of 20 cars have maladjusted
headlights, then
α = P[X ≥ 16 | p ≤ 0.5 n = 20] ≤ P[X ≥ 16 | p = 0.5 n = 20]
= 1 − P[X ≤ 15 | p = 0.5 n = 20] = 1 − 0.9941 = 0.0059.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 281 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Type I Errors
H 0 : θ ≤ θ0 , H1 : θ > θ 0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 282 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Type I Errors
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 283 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Type II Errors
In testing either of
H 0 : θ ≤ θ0 , H1 : θ > θ 0
H 0 : θ ≥ θ0 , H1 : θ < θ 0
H 0 : θ = θ0 , H1 : θ ̸= θ0
Type II errors are more tricky than Type I errors:
◮ A Type II error occurs if H0 is untrue but we still fail to reject H0 .
◮ We do not control β, the probability of committing a Type II error.
◮ β depends on the true value of the population parameter θ.
We also introduce the power of a test, which is he probability of rejecting
H0 when H1 is true. The power is given by 1 − β. Summarizing, we have
α = P[Type I error] = P[reject H0 | H0 true],
β = P[Type II error] = P[fail to reject H0 | H0 false],
Power = 1 − β = P[reject H0 | H0 false],
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 284 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 286 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 287 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 288 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 289 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 290 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Hypothesis Testing
2.3.15. Example. The burning rate of a rocket propellant is being studied.
Specifications require that the mean burning rate must be 40 cm/s.
Furthermore, suppose that we know that the standard deviation of the
burning rate is approximately σ = 2 cm/s. The experimenter decides to
specify a Type I error probability of α = 0.05 and he will base the test on
a random sample of size n = 25. The hypotheses we wish to test are
H0 : µ = 40 cm/s, H1 : µ ̸= 40 cm/s.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 291 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Hypothesis Testing
X −µ
−zα/2 ≤ Z ≤ zα/2 ⇔ −1.96 ≤ √ ≤ 1.96.
σ/ n
Twenty-five specimen are tested, and the sample mean burning rate
obtained is x = 41.25 cm/s. The value of the test statistic is
x − µ0 41.25 − 40
Z0 = √ = √ = 3.125.
σ/ n 2/ 25
We note that |Z0 | > 1.96, so it falls in the critical region and we can reject
H0 at the 5% level of significance.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 292 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Hypothesis Testing
2.3.16. Example. Continuing from the previous example, suppose that the
analyst is concerned about the probability of a Type II error if the true
mean burning rate is µ = 41 cm/s. We may use the following operating
characteristic curve (specific to α = 0.05) to find β:
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 293 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Hypothesis Testing
In this graph,
|µ − µ0 | 41 − 40 1
d := = = .
σ 2 2
Since in our example n = 25 we can read off β ≈ 0.30.
2.3.17. Example. Continuing from the previous example, suppose that the
analyst would like to design the test so that if the true mean burning rate
differs from 40 cm/s by more than 1 cm/s the test will detect this (i.e.,
reject H0 : µ = 40) with a high probability, say 0.90. This corresponds to
setting the power 1 − β of the test to 0.90, i.e., we want to achieve
β = 0.10.
We want to have β ≤ 0.1 if
|µ − µ0 | |µ − 40| 1
d= = ≥ .
σ 2 2
We see that the point (d, β) = (0.5, 0.1) is intersected by the OC curve
for n = 40 and that the curve remains below 0.1 for d > 1/2. Thus the
test should involve a sample size of n = 40 or more.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 294 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Significance Testing
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 295 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Significance Testing
For one-tailed test, the P-value is the area under the density curve to the
right or left of the observed statistic:
Significance Testing
2.3.18. Example. We test the hypothesis that a new car design increases
mileage.
◮ The population is the set of all cars with the new design;
◮ The random variable X is the mileage of the newly designed cars;
◮ The distribution of X is unknown;
◮ The population parameter µ is the mean of X .
We take a random sample of n = 36 automobiles. Our hypotheses are
H0 : µ ≤ 26 H1 : µ > 26
with the null value µ0 = 26. Currently, the mileage of cars has a standard
deviation of 5 miles and we assume this will also be true for the new
design if µ = µ0 .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 297 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Significance Testing
This is the P-value of the test. We may decide that it is sufficiently small
to reject H0 .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 298 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
or
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 299 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
X − µ0
T = √
S/ n
X − µ0
T = √
S/ n
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 302 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
which does not fall into the critical region, so there is insufficient evidence
to reject H0 at the 5% level of significance.
Note that the T -distribution may be used for XS/−µ
√ 0 when a sample is
n
obtained from a normal population. If a sample is obtained from a
non-normal population, care must be taken; for large to medium sample
sizes (n ≥ 25) it can be shown that violating the normality assumption
does not significantly change α and β. For small sample sizes, a T -test
cannot be used and an alternative (non-parametric) test must be
employed; such tests will be discussed later.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 303 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
These conventions emphasize the null value µ0 , for which a test statistic
with known distribution must be found.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 304 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 305 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 306 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Since the test is left-tailed, we reject H0 if this value is too small to have
occurred by chance. From the χ2 table we see that
Since the observed value lies between 14.6 and 18.3, the probability
P[χ219 ≤ 16.79] lies between 0.25 and 0.50. This probability is too large to
be able to reject H0 , so we cannot claim that σ < 1.5 mm.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 307 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 308 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 309 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 310 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 311 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Formally, we set
H0 : M = 2000,
H1 : M ̸= 2000,
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 312 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
While there are tables for checking the significance of such a result, we
can calculate it directly, because the signs of the differences are binomially
distributed. The probability of observing 6 or fewer negative signs in a
sample of 20 observations is
6 µ ¶
X 20 1 1
P[Q ≤ 6] = = 0.058.
r 2r 220−r
r =0
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 314 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 315 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 316 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
n(n + 1)
E[W ] =
4
and variance
n(n + 1)(2n + 1)
Var W = .
24
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 317 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 318 / 355
An Introduction to Statistical Methods Inferences on the Mean and Variance of a Distribution
Here
W+ = 1 + 2 + 3 + 4 + 5 + 6 + 11 + · · · + 19 + 20 = 150
and
|W− | = 7 + 8 + 9 + 10 + 12 + 14 = 60.
For our two-tailed test, we take W = min(60, 150) = 60. From the Table
VIII in Appendix A, with n = 20 observations we have the critical value of
52 for a two-tailed test with P = 0.05. Since W = 60 ̸< 52, we cannot
reject H0 at a 95% level of significance..
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 319 / 355
An Introduction to Statistical Methods Inferences on Proportions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 320 / 355
An Introduction to Statistical Methods Inferences on Proportions
Estimating Proportions
One of the (mathematically) simplest population parameters of general
interest is the proportion of members of a population with some trait.
Every member of the population is characterized as either having or not
having this trait. We describe this mathematically by defining the random
variable (
1 has trait,
X =
0 does not have trait.
The proportion of the members of the population having the trait is
1 X
N
# members wih trait
p= = xi
population size N
i=1
where N is the population size and xi is the value of the variable X for the
ith member of the population. Hence the proportion is equal to the mean
of X .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 321 / 355
An Introduction to Statistical Methods Inferences on Proportions
Estimating Proportions
It follows that if we take a random sample X1 , . . . , Xn of X , the sample
mean
1X
n
p̂ = X = Xi
n
i=1
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 322 / 355
An Introduction to Statistical Methods Inferences on Proportions
Estimating Proportions
We therefore obtain the following 100(1 − α)% confidence interval for p:
p
p̂ ± zα/2 p(1 − p)/n
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 323 / 355
An Introduction to Statistical Methods Inferences on Proportions
Estimating Proportions
or r r
0.16 · 0.84 0.16 · 0.84
0.16 − 1.96 ≤ p ≤ 0.16 + 1.96
75 75
which simplifies to 0.08 ≤ p ≤ 0.24.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 324 / 355
An Introduction to Statistical Methods Inferences on Proportions
2.4.2. Example. A new method of precoating fittings used in oil, break and
other fluid systems in heavy-duty trucks is being studied. How large a
sample is needed to estimate the proportions of fitting that leak to within
0.02 with 90% confidence?
Since no prior estimate is available, we take
2
z0.05 1.6452
n= = = 1692.
4d 2 4 · 0.022
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 326 / 355
An Introduction to Statistical Methods Inferences on Proportions
Hypothesis Testing
There a three types of hypotheses and tests for proportions. Let p) denote
the null value of a proportion p. Then we have
1. H0 : p = p0 , H1 : p > p0 (Right-tailed)
2. H0 : p = p0 , H1 : p < p0 (Left-tailed)
3. H0 : p = p0 , H1 : p ̸= p0 (Two-tailed)
For large sample sizes we use the following test statistic to test H0 : p = p0 :
p̂ − p0
Z=p .
p0 (1 − p0 )/n
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 327 / 355
An Introduction to Statistical Methods Inferences on Proportions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 329 / 355
An Introduction to Statistical Methods Inferences on Proportions
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 330 / 355
An Introduction to Statistical Methods Inferences on Proportions
Pooled Proportions
Most commonly we test against Ho : (p1 − p2 )0 = 0. Then we have the
hypotheses
1. H0 : p1 − p2 = 0, H 1 : p1 − p2 > 0 (Right-tailed)
2. H0 : p1 − p2 = 0, H 1 : p1 − p2 < 0 (Left-tailed)
3. H0 : p1 − p2 = 0, H1 : p1 − p2 ̸= 0 (Two-tailed)
Now if H0 is true, then p̂1 and pˆ2 are both estimators for the same
proportion p. Then the variance becomes
µ ¶
1 1
p(1 − p)/n1 + p(1 − p)/n2 = p(1 − p) +
n1 n2
and
p̂1 − p̂2
Z=r ³ ´
p(1 − p) n11 + 1
n2
Pooled Proportions
In estimating p, we now have a choice of p̂1 and p̂2 .It turns out that it is
best to take the weighted average
n1 p̂1 + n2 p̂2
p̂ =
n1 + n2
and to use the statistic
p̂1 − p̂2
Z=r ³ ´.
p̂(1 − p̂) n11 + 1
n2
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 332 / 355
An Introduction to Statistical Methods Inferences on Proportions
Pooled Proportions
2.4.4. Example. Many consumers think that automobiles built on Mondays
are more likely to have serious defects than those built on any other day of
the week. To support this theory, a random sample of 100 cars built on
Monday is selected and inspected. Of these, eight are found to have
serious defects. A random sample of 200 cars produced on other days
reveals 12 with serious defects. Do these data support the stated
contention?
We test
H 0 : p1 = p2 , H1 : p1 > p2
where p1 denotes the proportion of cars with serious defects produced on
Mondays.
Estimates for p1 and p2 are
p̂1 = 8/100 = 0.08, p̂2 = 12/200 = 0.06.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 333 / 355
An Introduction to Statistical Methods Inferences on Proportions
Pooled Proportions
From the standard normal table, we see that the probability of observing
this large or a larger value is 0.2546, so we shall not reject H0 .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 334 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 335 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
Comparing Variances
We have seen when estimating the mean of a single distribution that the
key lies in understanding the distribution of the variance. Therefore, we
first consider the comparison of variances.
We consider the two cases
1. H0 : σ12 = σ22 , H1 : σ12 > σ22 (right-tailed test)
2. H0 : σ12 = σ22 , H1 : σ12 ̸= σ22 (two-tailed test)
For comparing variances, we prefer to consider the quotient instead of the
difference of the estimators: If the sample variances are S12 and S22 , the
null hypothesis is true if S12 /S22 = 1, while we reject the null hypothesis if
the quotient is much larger than 1.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 337 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
Xγ21 /γ1
Fγ1 ,γ2 =
Xγ22 /γ2
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 338 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
Proof.
We know that (n1 − 1)S12 /σ12 and (n2 − 1)S22 /σ22 follows χ2 -distributions
with n1 − 1 and n2 − 1 degrees of freedom, respectively. Then
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 340 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
s12 4.02
2 = = 1.03.
s2 3.89
If α = 0.05, we find that the critical value for F7,7 is 3.787. Therefore,
there is not enough evidence to reject H0 .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 341 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
(X 1 − X 2 ) − (µ1 − µ2 )
p
σ12 /n1 + σ22 /n2
(X 1 − X 2 ) − (µ1 − µ2 )
p ,
σ 2 (1/n1 + 1/n2 )
and we are faced with the task of estimating σ 2 . We define the pooled
estimator
(n1 − 1)S12 + (n2 − 1)S22
Sp2 = .
n1 + n2 − 2
Then (n1 + n2 − 2)Sp2 /σ 2 will follow a χ2 -distribution with n1 + n2 − 2
degrees of freedom.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 342 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
Thus
(X 1 − X 2 ) − (µ1 − µ2 )
q
Sp2 (1/n1 + 1/n2 )
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 343 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 344 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 345 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
We can use the previous results for the testing of hypotheses on the
difference of means. If µ1 − µ2 = (µ1 − µ2 )0 (the null value of the
difference of means), the statistic
X 1 − X 2 − (µ1 − µ2 )0
Tn1 +n2 −2 = q
Sp2 (1/n1 + 1/n2 )
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 346 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
We can then analyze D using the methods for the mean of a single
random variable.
Hypothesis tests for paired data are called paired T -tests.
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 349 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
If the sample sizes are small, the variances unequal, and/or the populations
non-normal, the T -tests may not yield good results. In such situations, a
non-parametric test is available that is nearly as good as a T -test.
The Wilcoxon rank-sum test tests two random variables X and Y for
equality. However, it is especially sensitive to differences in location. Hence
we usually state the hypotheses in terms of the medians MX and MY :
◮ H0 : MX = MY , H1 : MX > MY (right-tailed),
◮ H0 : MX = MY , H1 : MX < MY (left-tailed),
◮ H0 : MX = MY , H1 : MX ̸= MY (two-tailed).
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 350 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 351 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
H0 : M1 = M2 , H1 : M1 ̸= M2 .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 352 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 353 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
The sums of the ranks are W1 = 99 for alloy 1, and W2 = 111 for alloy 2.
By Table X in Appendix A, the critical values for α = 0.05 in a two-tailed
test are 79 and 131. Since neither sum of ranks is outside the interval
[79, 131], we cannot reject H0 .
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 354 / 355
An Introduction to Statistical Methods Comparing Two Means and Two Variances
Dr. Hohberger (UM-SJTU JI) Probabilistic Methods in Engineering Spring 2009 355 / 355