Nonparametric Statistics: Significance of Computed

You might also like

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 1

Nonparametric Statistics

Parametric statistics serve as estimates of the corresponding parameters. Their computations require the use of
precomputed statistics as estimates of parameters. Moreover, they are interpreted with reference to specific population
distributions of the variables such as the normal and t distributions. Thus, they cannot be used for too small samples,
nominal, ordinal and discrete variables, and non-normal distributions.

Nonparametric (distribution-free) statistics (i) require very few assumptions, (ii) do not require normal distributions of the
variables in the population, (iii) do not use any precomputed statistic as an estimate of parameter in the computation, (iv)
can be used for even very small samples and (v) are computed by much simpler methods ; but (vi) their powers are lower
than their parametric counterparts so long as the assumptions for the latter are fulfilled.

Chi Square Tests


Nonparametric chi square (𝜒 2 ) test explores the significance of deviation of an experimentally observed frequency
distribution from a proposed frequency distribution and, therefore, constitutes an analysis of frequencies. It (i) requires no
assumption for normality of the population distribution of variable(s), (ii) uses no precomputed statistic as an estimate of
parameter in its computation, (iii) is applicable to very small samples, and (iv) can be used also for discrete, nominal or
ordinal variables.

𝜒 2 is the sum of the ratios of squared deviations of observed frequencies (𝑓𝑜 ) from the frequencies (𝑓𝑒 ) expected from a
given distribution, and the respective 𝑓𝑒 values.

(𝑓𝑜 − 𝑓𝑒 )2
𝜒2 = Σ
𝑓𝑒

𝜒 2 has positively skewed probability distributions (Fig. 9.1) depending


on its df. The area under the 𝜒 2 distribution curve for the given df is
taken as 1.00. Each 𝜒 2 distribution is asymptotic to the abscissa in its
right tail — its right tail reaches the abscissa at +∞ only. 𝜒 2 cannot have
negative values. When df ≃ ∞, the distribution is very little skewed;
indeed, it approaches almost a bilateral symmetry at df > 30.

With the fall in df positive skewness increases progressively. At df = 1,


the distribution is L-shaped or reverse J-shaped, asymptotic to both X
and Y axes. Thus the probability of getting a given 𝜒 2 by mere chance
due to random sampling equals the fractional area beyond the ordinate
at the given 𝜒 2 in the right tail of the distribution of 𝜒 2 values with the
specified df.

Significance of computed 𝝌𝟐 :

The 𝐻𝑜 contends that the computed 𝜒 2 is not significant and has resulted from mere chances of random sampling. To test
this 𝐻𝑜 by a two-tail test, the computed 𝜒 2 is compared to the critical 𝜒 2 with the computed df and for the chosen level of
significance (α). If the computed 𝜒 2 exceeds or equals the critical 𝜒 2 , the probability P of obtaining the computed 𝜒 2 by
chance is respectively lower than and equal to α and may be considered too low (P ≤ α). The 𝐻𝑜 is then rejected and the
computed 𝜒 2 is considered significant – the variables are then considered to have significant association with each other.
But if the computed 𝜒 2 is lower than the critical 𝜒 2 , P exceeds α and is too high to justify the rejection of 𝐻𝑜 (P > α); the
computed 𝜒 2 is then not significant – the variables have no significant association with each other.

For a one-tail 𝜒 2 test, a for a given critical 𝜒 2 is half that of an identical critical 𝜒 2 for two-tail tests. In other words, the one-
tail critical 𝜒 2 for a given significance level is identical with the two-tail critical 𝜒 2 for double that significance level. Thus,
the one-tail critical 𝜒 2 (df=4) for 0.05 level amounts to 7.78 which is identical with the two-tail critical 𝜒 2 (df=4) for 0.10
level.

You might also like