Professional Documents
Culture Documents
Measures of Variability
Measures of Variability
In module #3, the different measures of central tendency (mean, median & mode) and point measures (percentile,
decile & quartile) were discussed. However, these measures do not give us the total picture of the samples we have
measured. They do not give us information on how the data are scattered or provide a mass description of the data. It
is therefore necessary to have measures of variation or dispersion to describe better the characteristics of a set of
data.
The variation of one measurement from another is an enduring characteristic of any sample of measurements. But if a
sample is only one which involves only one score, there is no variation. Variation occurs in a study if it involves many
subjects with different scores or sets of observations.
While measures of central tendency/location describe the typical or representative values, measures of
variability/variation describe the degree of dispersion, scatter or spread of scores in a distribution
THE RANGE
The simplest and easiest measure of variability; easily obtained and readily understood. However, since its
magnitude depends on the extreme values of the set of data (the highest and the lowest), it is the most
unreliable.
Two types: exclusive and inclusive range. The exclusive range (ER) is the difference between the largest and
the smallest scores in a given distribution; whereas the inclusive range (IR) is the difference between the
exact upper limit of the class interval containing the largest score and the exact lower limit of the class interval
containing the smallest score.
EXAMPLE:
Highest Score = 10 Lowest Score = 3
THEN:
ER= 10-3 IR = 10.5 – 2.5
=7 =8
NOTICE that the IR is 1 more than the ER. The IR is therefore equal to the highest score minus the
lowest score plus one.
The simplest method of taking into account the variation of all the items in a series from the point of central tendency
which may either be the mode, the median or the mean.
The sum of the absolute deviations of the arithmetic mean divided by the number of cases.
LIMITATIONS:
1. Since it is based upon all the deviations, it may be inflated by a single extreme value.
2. The use of the absolute values of the deviations in its computation accounts for its infrequent use in statistical
theory.
Like the average deviation, the variance and the standard deviation are based on the deviations of all the scores in
the series. Unlike the average deviation, however, which is reckoned either from the mean, mode or median, the
variance and the standard deviation are always computed from the mean.
Another difference is that the signs of the deviations are disregarded in computing these two measures of variation.
The negative signs are eliminated by squaring the deviations. The squared deviations are added then divided by the
number of cases n. The result is called the variance. The square root of the variance, on the other hand, is known as
the standard deviation which is expressed not in terms of square units as the variance, but in terms of the linear unit
of measurement of the given quantity. For example, in terms of the variable length, variance maybe expressed in
square feet or square meters, while standard deviation will be in feet or meters.
The choice of an absolute measure of variation to use depends on the following considerations:
COEFFICIENT OF VARIATION
Expresses the standard deviation as a percentage of the mean.
Can be used to compare the variability of two or more distributions even when the observations are
expressed in different units of measurement.
STANDARD SCORES
Tells the relative location of a particular raw score with regard to the mean of all the scores in a
series.
Raw scores are scores or measurements which are expressed in the form in which they were
originally obtained, like test scores and weight and height measurements.
Standard scores are transformed raw scores. The transformation is made by subtracting the mean
from the given raw score and dividing the difference by the standard deviation of the series of raw
scores.
Standard scores, therefore, are expressed in terms of standard deviation units from the mean.
The mean of standard scores is ZERO. Hence, a positive standard score indicates that the
transformed raw score is above or higher than the mean, while a negative standard score shows
that the given raw score is below or lower than the mean.
The formula for transforming a raw score to a standard score, represented by z, is:
Z= x - x
S
Standard scores are usually used to compare observations in two or more different standard
deviations. They are also used to compare raw scores from two or more distributions which are
expressed in different measurement units like final grades and achievement test scores.
T-scores are another type of standard scores. A z score may be converted to a t score with the
formula:
T= 10z + 50
If z scores are converted to t scores, then one does not have to manipulate with negative values
since t scores are all positive.
SKEWNESS
The asymmetrical distribution of a set of data in a scale with respect to its arithmetic mean. A
distribution can be positively or negatively skewed.
A positively skewed distribution tapers more to the right, which means concentration of the values
is below the arithmetic mean.
A negatively skewed distribution tapers more to the left, which means concentration of values is
above the arithmetic mean.
When the mean is larger than the median, the distribution is skewed to the right. When the mean
is lower than the median, then the distribution is skewed to the left.
In a normal distribution, sk=0. If sk has a negative value, then the distribution is skewed to the
left. If sk has a positive value, the distribution is skewed to the right. The greater the value of sk departs from zero, the
more skewed or asymmetrical is the distribution. The nearer the distribution is to normal, the closer to zero is the value of
sk.
KURTOSIS
A measure of the degree of peakedness or flatness of a distribution. When the distribution is normal, ku=3. If ku is
less than 3, the distribution is platykurtic or less peaked than the normal curve. If ku is greater than 3, the distribution is
leptokurtic or more peaked than the normal curve.