Professional Documents
Culture Documents
INFERENCE
INFERENCE
Structure
19.0 Objectives
19.1 Introduction
19.2 Theory of Estimation
19.2.1 Parameter Space
19.3 Characteristics of Estimators
19.3.1 Consistency
19.3.2 Unbiasedness
! 19.3.3 Efficiency
19.3.3.1 Most Efficient Estimator
19.3.3.2 Minimum Variance Unbiased Estimator
19.3.4 Sufficiency
19.4 Cramer: Rao Inequality
19.4.1 MVUE and Blackwellisation
19.4.2 Rao -Blackwell Theorem
19.5 Test of Significance
19.5.1 Null Hypothesis
19.5.2 Alternative Hypothesis
19.5.3 Critical Region and Level of Significance
19.5.4 Confidence Interval and Confidence Limits
19.5.5 One: Tailed and Two: Tailed Test
19.5.6 Critical Values and Significant Values
19.6 Type I and Type I1 Error
19.7 Power of the Test
19.8 Optimum Test under Different Situations
19.8.1 Most Powerful Test
19.8.2 Uniformly Most Powerful Test
19.9 Test Procedure under Normality Assumption
19.9.1 Problems Regarding the Univariate Normal Distribution
19.9.2 Comparison of Two Univariate Normal Distribution
19.9.3 Problems Relating to a Bivariate Normal Distribution
19.10 LetUs Sum Up
1 9.1 1 Key Words
19.12 Some Useful Books
19.13 Answer or Hints to Check Your Progress
19.14 Exercises
19.0 OBJECTIVES
After going through this unit, which explains the concepts of estimation
theory and hypothesis testing, you will be able to answer the following:
How characteristics of any population can be inferred on the basis of
analysing the sample drawn from that population;
The likeliness of any characteristic of a population on the basis of
analysing the sample drawn from that population;
What should be the characteristics of an estimator?
What are the test criteria under different situations?
-
Statistical Methods 11
19.1 INTRODUCTION 1
The object of sampling is to study the features of the population on the basis
of sample observations. A carefully selected sample is expected to reveal
these features, and hence we shall infer about the population from a statistical
analysis of the sample. The process is known as 'statistical inference'.
I
There are two types of problems. First, we may have no information at all
about some characteristics of the population, especially the values of the
parameters involved in the distribution, and it is required to obtain estimates
of these parameters. This is the problem of 'estimation'. Secondly, some
I'
information or hypothetical values of the parameters may be available, and it
is required test how far the hypothesis is tenable in the light of the information
provided by the sample. This is the problem of 'hypothesis testing' or 'test of
significance'. -1
19.2 THEORY OF ESTIMATION
Suppose, we have a random sample XI,x2, ... , x, on a variable x, whose
distribution in the population involves an unknown parameter 0. It is required
to find an estimate of 0 on the basis of sample values. The estimation is done
in two different ways: (i) Point estimation, and (ii) Interval estimation.
In 'point estimation', the estimated value is given by a single quantity, which
is a function of sample observations (i.e., statistic). This function is called
'estimator'. .
In 'interval estimation', an interval within which a parameter is expected to lie
is given by using two quantities based on sample values. This is known as
'confidence interval', and the quantities, which are used to specify the
interval, are known as 'confidence limits'. Since our basic objective is to
estimate the parameter associated with the sample conservations, so before
going into further details, let us discuss elaborately about the notion of
parameter space.
19.2.1 Parameter Space
Let us consider a random variable (r. v) x with p. d. f. f(x, 0). In most common
I
application, though not always, the functional form of population distribution
is assumed to be known except for the value of some unknown parameter(s) 8,
which may take any value on a set 0 . This is expressed by writing the p. d. f.
in the form f(x, 0), 0 E O. The set 0 , which is the set of all possible values of
0, is called the 'parameter space'. Such a situation gives rise not to one
probability distributions but a family of probability distribution which we
write as [f(x, 0), 0 E a}.For example, if X- N (p, 02), then the parameter
space@= {(p, 0'):-co<p<co;0 <o<co)
I, 2, ... , k.}.
Let us consider a random sample xl, x2, .. . , x,. of size 'n' from a population, I
with probability function f(x; 0,, e2, ... , €Ik),where 81, 02, ... , Ok are the
unknown population parameters. There will then always be an infinite number
of functions of sample values, called statistic, which may be proposed as
estimates of one or more of the parameters.
Statistical Inference
Evidently, the best estimates would be one that falls nearest to the true value
of the parameter to be estimated. In other words, the statistic whose
distribution concentrates as closely as possible near the true value of the
parameter may be regarded the best estimate. Hence, the basic problem of
estimation in the above case can be formulated as follows:
We wish to determine the functions of the sample observations:
n n
T I = 8 ] ( x ~X72, ... xn), T2= Q2(x1, X2, -xn), ... Tk= Q k (XI,X2,
0 . xn), 0 . .
such that their distribution is concentrated as closely as possible near the true
value of the parameter. The estimating functions are then referred to as
'estimator'.
L= fi
I=1
f (XI$3)
= g 1 ( t , , 0 ) . k ( x 1 , x *..., ,XI,)
where gl(tl, 0) is the p.d.f. of statistic tl and k(xl, xz, ... , x,) is a function
of sample observations only, independent of 0.
Note that this method requires working out of the p.d.f. (p.m.f.) of the
statistic tl(xl, x2, ... , x,), which is not always easy.
Stntistical Methods - 11
Check Your Progress 1
1) Discuss the meaning of point estimation and interval estimation.
......................................................................................
2) List the characteristics of a good estimator.
3) xl, x2, .... x, is a random sample from a normal population N (p, 1).
Show that F =-x
n
1 "
;=,
xz , is an unbiased estimator of p2 + 1.
......................................................................................
4) A random sample (XI, X2, X3, X4, X5) of size 5 is drawn from a normal
population with unknown mean p. Consider the following estimators t o
estimate p:
i) ti =(XI + X2+ X3+ X4+ X5) 15; and (ii) t2 = (2x1 + X2+ hX3)13.
Find h. Are tl and t2 unbiased? State with giving reasons, the estimator
which is best among tl and t2.
i =l
, is sufficient for 0.
Blackwellisation.
If in addition, the sufficient statistic T is also complete, then the estimator
cp(T) discussed above will not only be an improved estimator over U but also
the 'best (unique)' estimator.
If n is small (usually less than 30), then the sampling distribution of the test
statistic Z will not be normal and in that case we cannot use the above
significant values, which have been obtained from normal probability curves.
Procedure for Testing of Hypothesis:
We now summarise below the various steps in testing of a statistical
hypothesis in a systematic manner.
1) Null Hypothesis: Set up the null hypothesis Ho.
2) Alternative Hypothesis: Set up the alternative hypothesis H I . This will
enable us to decide whether we have to use a single-tailed (right or left)
test or two-tailed test.
3) Level of Significance: Choose the appropriate level of significance ( a )
depending on the reliability of the estimates and permissible risk. This is
to be decided before sample is drawn, is., a is fixed in advance.
Compute the test statistic
......................................................................................
2) What do you mean by test of significance?
......................................................................................
3) What is the purpose of hypothesis testing?
The main objective in sampling theory is to draw valid inferences about the
population parameters on the basis of the sample results. In practice, we
decide to accept or reject the lot after examining the examining the sample
from it. As such, we are liable to commit the following two types of errors:
Type I Error: Reject Howhen it is true.
Type I1 Error: Accept Howhen it is wrong, i.e., accept Howhen H I is true.
If we write, P [Reject Howhen it is true] = P [Reject Ho / Ho] = a
and P [Accept Howhen it is wrong] = P [Accept Ho I HI] = p
then a and 8 are called the sizes of type I error and type I1 error respectively.
In practice, type I error amounts to rejecting a lot when it is good and type I1
error may be regarded as accepting the lot when it is bad.
Thus, P [Reject a lot when it is good] = a
and P [Accept a lot when it is bad] = P
where a and p are referred to as 'Producer's risk' and Consumer's risk'
respectively.
--
The. probability of type I error is necessary for constructing a test of Statisttal Inference
1 "
s" the sample variance of x: s" = - x (x, -I)' . The distinction between s2
n -1
and sY2is to be noted. In the divisor is (n-1), which makes it as unbiased
estimator ofo2.
n
1 1 I1
1 1 2 0'
= -{ Z var(xi) - n var(ii)} = -{no -n-}=02
n-1 , n-1 n
Case I: ,u unknown, o known
Here we may be required to test the null hypothesis Ho: p = [LO.It has already
been shown that the test procedure for HO in this case is based on the
statistic&(^- p,)/o, which is distributed as a normal deviate ( r ) under this
hypothesis.
a) For the alternative H: p > b , Ho is rejected if for the given sample
z > r, (and is accepted otherwise)
c) For the alternative H: p#b, HOis rejected if for the given sample I T ) >r,,,?.
Statistical Inference
In each case, a denotes the chosen level of significance.
As regards the problem of interval estimation of p, it has been shown that the
limits (Y- r,,, .o/&) and ( P + r,,, .o/&), computed for the given sample,
are the confidence limits for with confidence coefficient (1 - a ) .
Case ZZ: p known, o unknown
Here one may be interested in testing a hypothesis regarding o or in
estimating o .
Hence, )I
, being the sum of squares of n independent normal
a) For the alternative N: o>o,, No is rejected in case for the given sample
2 2
X > X,., .
b) Fo;. the alternative N: o < o,, Ha is rejected if for the given sample x',
XI-,.,
c) For the alternative H: o f o , , Ho is rejected if for the given sampleX',
2 2 2
X I - ~ : Z Or'
,~ X >Xa/~,n .
estimate, s' =
\il z
-- (xi -
n - 1 ,=,--
.
s'
The 100(1 - a ) % confidence limits to p will, therefore, be (K - t,,,, ,,.,)---
A
S'
and (ST+ to,,, ,-,)-, these being computed from the given sample.
&
'The procedure has been called as Student's t-test.
In this case, we may have also the problem of testing Ho:o = o, or the problem
of obtaining confidence limits t o o . From what has been said above, it is clear
I,
Z ( x , -q2
that '=I - (n - 1) s ' ~is, under the hypothesis Ho, a X2 with df = (n-I).
6; 0;
This provides us with test for Ho. The value of this x', computed from the
given sample, is compared with X,2 ,,-, or x 2,-,, ,-, , according as the alternative is
H: o > o , o r H o<o,.
For the alternative H: o f o , , on the other hand, the computed value is to be
compared with both x:-,,, ,,-, and&, ,,-, , Ho being rejected if the computed
value is smaller than the former or exceeds the latter value.
) e., j.[
(n-1)s''
~:/2.n-l
50 5
(n-l)st2
xL!?,"-I I = 1 -(-j
Statistical Inference
i
Check Your Progress 3
1) Which is wider, a 95% or a 99% confidence interval?
2) When you construct a 95% confidence interval, what are you 95%
confident about?
......................................................................................
3) When computing a confidence interval, when do you use t and when do
you use z?
......................................................................................
4) What Greek letters are used to represent the Type I and I1 error rates.
......................................................................................
......................................................................................
5) What levels are conventionally used for significance testing?
......................................................................................
7) Distinguish between probability value and significance level?
Statistical Methods - II
......................................................................................
......................................................................................
8) The following are 12 determinations of the melting point of a compound
(in degrees centigrade) made by an analyst, the true melting point being
165'~. would you conclude fiom these data that her determination is
free from bias?
164.4, 161.4, 169.7, 162.2, 163.9, 168.5, 162.1, 163.4, 160.9, 162.9.
160.8, 167.7.
xIj 1 n, and
-
- x,)' I (n, - 1) are the mean and standard deviation of x in the
"1 "2
set confidence limit; to this ratio. Since (x,, - p,)' /of and
j=l
"2
respectively, P
I
Z ( x , , - P1l2 Inlo;
J=
n.
I
1
Fu12;n2,nl
I
I
Under the hypothesis Ho: o , / 0 2 = F,therefore,
J= I
n2 x-is an
Z(x2j - ~ 2 ) ' f a ;
502
j=l
F with df = (nl, n2). This provides a test for Ho. When the altemative is H: a,1
o, > b, HOis to be rejected if for the given samples F > F a ; ,I, ,2.
If the alternative is H. o,/ o,< to, HO is to be rejected if for the given samples
F<F~.a;ni,nz,i.e., if l / F > F a , n ~ , n l .
Lastly, when the alternative is H: o , / a,# 50, HO is to be rejected if the
samples in hand give either F < F1. a n , , I , d ,i.e., 1/F> Fa12; n2, n,, or
F > Fan;n~,nl.
-
Statistical Methods I1
The commonest form of the null hypothesis will be Ho: o , = 0 2 for which to=
"I
C ('IJ - P I l2
I , and here F = :'
-~ 2 1 ~ ~ ~ 2
J= I
I 1
"I
- l2 Inlo?
1 <
-
,=, PI
I =]-a
"2
n2, nl Fa/2,nl,n2
c ( x 2 J - ~2 l2 in2(J:
J= I
I.e.,
I 1
"I "I
1
C ( x , - PIl2/"I o2
1- ,=I
C(xiJ - PI 1' 11'
,=I
"2
sL<- "1 =I-a
'012, ni?, n t 0: Fr/2,nl,n2
(2' , - ~2 l2In2 C(x2, - P212 In2
P J=I I= I
o2
The confidence limits to -I- (with confidence coefficient 1 - a ) will, therefore,
02'
"I nl
Z@lJ - PlI2/n1 C(x1, - P1I21n,
1 ,=I 1 ,=I
be and
2 "1
C(xy -
'012, nl,
J= I
'012, "2, n l
Z@ZJ- ~
J= I
21%
) ~
- "
x 2 )+ (ta,2,n,tnl-2)
(-X, - - s . Obviously, in both cases we are using
z(xlj -~ , ~ / -n i)o:
, t 2
j -
- S ,x-is
, 0: an F with (n, - 1, n2 - 1) degrees
"1
1(x2, - z2>,/(n2 - 1) o: s2 6:
j
of freedom.
For testing Ho: o , / o,= 50, we will use the F statistic, but now F =
si2 1
-;-xl with (n, - 1, n2 - 1) degrees of freedom. The confidence limits to
s2' 5 0
1 si2 1 s;~
o o now will be -' 2 and -, . The
'012 nl-1, n2-1 2' '012; n2-1, n l - I '2
......................................................................................
3) State the effect on the probability of a Type I and of a Type I1 error of:
a) the difference between population means
b) the variance
c) the sample size
d) the significance level
4) The following data are the lives in hours of two batches of electric bulbs.
Test whether there is a significant difference between the batches in
respect of average length of life.
Batch I: 1505, 1556, 180 1, 1629, 1644, 1607, 1825, 1748.
Batch 11: 1799, 1618, 1604, 1655, 1708, 1675, 1728.
I
Here the s a m ~ l ecorrelation coefficient is r =
where Xand Yare the sample means. When p = 0, the sampling distribution of
I (1 - r 2 ) ( n - 4 ) / 2
r assumes a s i m ~ l eform f ( r ) = , and in that case
-(2' 2 1
r Jn-21 4 s can be shown to be distributed as a t with df = (n - 2).
This fact provides us with a test for Ho: p = 0. As to the general hypothesis Ha:
I'
p = po, an exact test becomes difficult, because for p # 0 the sample
correlation has a complicated sampling distribution. For moderately large n,
there is an approximate test, which we have not discussed presently.
Problemv regarding the difference between p, and p,,:
Information regarding the difference between the means, px and py,may be of
some importance when x and y are variables measured in the same units.
To begin with, we note that if we take a new variable, z = x - y, then this z,
being a linear function of normal variables, is itself normally distributed with
mean CL, = II,- py and variance og=o: + o; - 2 poxoy
It will follow, from what we have said in the section for univariate normal
1
distribution, that if we put z, = xi - yi, F = 7,
z;/ n and sL2 = -7 (z; - F ) ,~
PId i x
I ruv I
I
sto/2,n-2 = 1 - a
or, P[r:v
I
(n -2 2) 5 t:12,n-2 = 1- a
1 - ruv
By solving the equation r"?,,(n - 2)= t:,,,,-, (1 - r,?,,) or, say, v (6) = 0 for the
unknown ration 5 = a , l o , , two roots will be obtained. In case, the roots, say
5, and 52, are real (5, < t2), this will be the required confidence limits for 5
with confidence coefficient (1 - a ) .
Again, v (6) may be either a convex or concave function. In the former case,
we shall say 61 5 5 5 52, while in the latter we shall say 0 5 5 (1 or t25 5 1 oo.
But the roots may as well be imaginary, in which case we shall say that for the
given sample 100(1 - a ) % confidence limits do not exist.
Check Your Progress 5
1) The correlation coefficient between nasal length and stature for a group
of 20 Indian adult males was found to be 0.203. Test whether there is
any correlation between the characters in the population.
......................................................................................
......................................................................................
......................................................................................
2) a) What proportion of a normal distribution is within one standard
deviation of the mean? b) What proportion is more than 1.8 standard
deviations from the mean? c) What proportion is between I and 1.5
standard deviations above the mean?
......................................................................................
......................................................................................
......................................................................................
3) A test is normally distributed with a mean of 40 and a standard deviation
of 7. a) What score would be needed to be in the 85'h percentile? b) .
What score would be needed to be in the 22"d percentile?
......................................................................................
......................................................................................
......................................................................................
4) Assume a normal distribution with a mean of 90 and a standard
deviation of 7. What limits would include the middle 65% of the cases.
......................................................................................
......................................................................................
......................................................................................
Statistical Inference
19.10 LET US SUM UP
1 In this unit, we have learnt how by using the theory of estimation and test of
1 significance, an estimator can be analysed and how sample observations can
be tested for any statistical claim. The unit shows the way of testing various
real life problems through using statistical techniques. The basic concepts of
hypothesis testing as well as of estimation theory are also made clear.
it 1911 KEY WORDS
/ Alternative Hypothesis: Any hypothesis, which is complementary to the null
hypothesis, is called an alternative hypothesis, usually denoted by HI.
I Confidence Interval and Confidence Limits: If we choose once for all some
C
small value of a (5% or 1%) i.e., level of significance and then determine two
constants say, cl and c2 such that P [cl < 8 < c2]=l - a , where 8 is the
unknown parameter then, the quantities c1 and c2, SO determined, are known as
the 'confidence limits' and the interval [cl, c2] within which the unknown
value of the population parameter is expected to lie, is called the 'confidence
interval' and (1 - a ) is called the 'confidence coefficient'.
Consistency: T, is a consistent estimator of y(8) if for every E > 0, q > 0,
there exists a positive integer n >_ m ( ~q)
, such that
P [IT, - y(8)I < E ] + 1 as n + m = > P [ITn - y(8)I < E] > 1 - q ; t l n L m
where m is some very large value of n.
Cramer-Rao Inequality: If t is an unbiased estimator ofy(8), a hnction of
parameter 8, then
2
condition.
Null Hypothesis: A definite hypothesis of no difference is called 'null
hypothesis' and usually denoted by Ho.
One -Tailed and Two - Tailed Tests: A test of any statistical hypothesis
where the alternative hypothesis is one-tailed (right-tailed or left-tailed) is
called a 'one-tailed test'. For example, a test for testing the mean of a
population Ho: p = p~ against the alternative hypothesis: HI: p > p~ (right-
tailed) or HI: p < ~b (left-tailed), is a 'one-tailed test'.
Let us consider a test of statistical hypothesis where the alternative hypothesis
is two-tailed such as: Ho: p = b , against the alternative hypothesis HI: p f
(p > wand p < w) is known as 'two-tailed test'.
Parameter Space: Let us consider a random variable x with p. d. f. f(x, 0).
The p. d. f. of x can be written in the form f(x, 0), 0 E 0 . The set 0 , which is
the set of all possible values of 0, is called the 'parameter space'.
Power of the Test: The power of the test can be defined as
Power = 1 - Probability of Type I1 error
= Probability of rejecting Howhen HI is true.
Sufficiency: If T, = T (XI,x2, . . . , x,) is an estimator of a parameter 0, based
on a sample X I ,x2, ... , x, of size 'n' from the population with density f(x, 0)
such that the conditional distribution of xl, x2, ... , X, given T, is independent
of 0, then T is sufficient estimator for 0.
Type I and Type I1 Error: Type I Error: Rejecting the null hypothesis Ho
when it is true.
Type I1 Error: Accepting the null hypothesis HOwhen it is wrong, i.e., accept
Howhen HI is true.
Unbiasedness: A statistic T, = T (xi, x2, ... , x,), is said to be an unbiased
estimator of y(0) if E (T,) = y(0), for all 0 E 0
Uniformly Most Powerful Test: The critical region W is called uniformly
most powerhl critical region of size a(and the corresponding test as
uniformly most powerful test of level a ) for testing Ho: 0 = 00 against HI: 0 f
00. if P (XEW I Ho) = a and P (XEW I HI) > P (XEWI I HI) for all 0 f 00,
whatever the other critical region Wt satisfying the first condition may be.
I ..
4) Solution: We are given E (Xi) = p, var (xi2) = a 2 , (say);
COV(Xi, Yi) = 0 (i # j = 1, 2, ... , n)
I-
1
i) E (t,) = - Z E ( ~ , ) = ( 1 / 5 ) . 5=~p=> tl is an unbiasedestimatur of
5 ,=,
CL-
ii) E(t3)=p=>hp=O=>h=0
v (tl) = [v(X[) + v(X2) + v(X3)+ v(X4)+ v(X5)] / 25 = 0'15
v (t2) = XI) + v(X2)] 1 4 + v(X3) = 3 a 2 / 2
P
v (t3) = [4v(XI)+ v(X2)I / 9 = 5 a219 (since h = 0)
1 Since v (t,) is the least, tl is the best estimator (in the sense of least
t variance) of p.
d f = n , + n2 - 2. Andto02~,
13 = 2.160,to005,13=3.012.
19.13 EXERCISES
1) If T is an unbiased estimator of 8, show that T~ is a biased estimator for
e2. .
[Hint: Find var (T) = E ( T ~ )- €?. Since E ( T ~ )# e2, T~ is a biased
estimator for e2.]
Statistical Inference
2) XI, X2, and X3 is a random sample of size 3 from a population with
mean value p and varianceo2, T I , Tz, T3 are the estimators used to
estimate mean value p, where
TI = X i + X z - X3; T2 = 2XI - 4x2 + 3x3; andT3 = (?,XI + X 2 + X 3 ) / 3 .
i) Are TI and TZunbiased estimators?
ii) Find the value of h such that T3 is unbiased estimator of p.
I
iii) With this value of h is T3 a consistent estimator?
1
/ iv) Which is the best estimator?
i
[Hint: Follow Check Your Progress 21
3). Let XI, X2, . . . , X, be a random sample from a Cauchy population:
i 1 I
f (x, 8) =-. - m < x < m , - m<e<cO.
n 1 + ( x - 8)2 '
Examine if there exist sufficient estimate for 8.
n
[Hint: L(x, 0) = n f (xi, 0)
i= 1
Give two limits between which the mean weight at birth for all such
babies is likely to lie.
[Hint: Let us denote by x the variable: weight at birth per baby. Our
problem here is then to find, on the basis of the given sample of 15
babies, confidence limits for the population mean of x. We shall assume
(a) that in population x is normally distributed (with a mean p and
standard deviationo, both of which are unknown) and '(b) that the given
observations form a random sample from the distribution.
Under these assumptions, the 100(1 - a ) % confidence limits to p will
C ( x i - q2
assumption the test is given byx2 = '='
0:
1
Here, 'Z= 2.5, s:= 3.171, t = 2.493 and tabulated value is t0.05, = 1.833
and tool,9 = 2.821. The observed value is thus significant at the 5% but
insignificant at the 1% level of significance. If we choose the 5% level,
i then the null hypothesis should be rejected and we should say that the
change of diet results in a gain in average weight.
NOTES