Professional Documents
Culture Documents
ABPR1103
ABPR1103
ABPR1103
ABPR1103
Statistics for Psychology
INTRODUCTION
ABPR1103 Statistics for Psychology is one of the courses offered by the Faculty of
Applied Social Sciences at Open University Malaysia (OUM). This course is
worth 3 credit hours and should be covered over 8 to 15 weeks.
COURSE AUDIENCE
This course is offered to all learners taking the Bachelor of Psychology with
programme.
STUDY SCHEDULE
It is a standard OUM practice that learners accumulate 40 study hours for every
credit hour. As such, for a three-credit hour course, you are expected to spend
120 study hours. Table 1 gives an estimation of how the 120 study hours could be
accumulated.
Study
Study Activities
Hours
Briefly go through the course content and participate in initial discussions 2
Study the module 60
Attend 3 to 5 tutorial sessions 8
Online participation 12
Revision 18
Assignment(s), Test(s) and Examination(s) 20
TOTAL STUDY HOURS ACCUMULATED 120
COURSE OUTCOMES
By the end of this course, you should be able to:
COURSE SYNOPSIS
This course is divided into 10 topics. The synopsis for each topic can be listed as
follows:
Topic 4 shows the next step of statistical method ă summarise data. We will
discuss the measures of central tendency, which include the mean, mode and
median in statistics. It also discusses organising data by using quartiles, deciles
and percentiles.
Topic 6 explains the normal distribution or normal curve. It also shows us the
way to compute probability, and discusses sample and population.
Learning Outcomes: This section refers to what you should achieve after you
have completely covered a topic. As you go through each topic, you should
frequently refer to these learning outcomes. By doing this, you can continuously
gauge your understanding of the topic.
Summary: You will find this component at the end of each topic. This component
helps you to recap the whole topic. By going through the summary, you should
be able to gauge your knowledge retention level. Should you find points in the
summary that you do not fully understand, it would be a good idea for you to
revisit the details in the module.
Key Terms: This component can be found at the end of each topic. You should go
through this component to remind yourself of important terms or jargon used
throughout the module. Should you find terms here that you are not able to
explain, you should look for the terms in the module.
PRIOR KNOWLEDGE
No prior knowledge required.
ASSESSMENT METHOD
Please refer to myINSPIRE.
REFERENCES
Aron, A., Aron, E. N., & Coups, E. J. (2009). Statistics for Psychology (5th ed.).
Upper Saddle River, NJ: Pearson.
Urdan, T. C., & Urdan, T. C. (2005). Statistics in Plain English. Lawrence Erlbaum
Associates, Inc.
INTRODUCTION
Many learners take psychology degrees because they are interested in or curious
about peopleÊs behaviours. However, not many learners know that they have to
study research methods and statistics in detail, which seem scary for learners
who dislike or are poor in mathematics.
ACTIVITY 1.1
Post your answer to the above question in the myINSPIRE forum and
respond to some of the postings by your coursemates.
In psychology, we are aware that our observations of the world can never be
totally accurate but have a certain level of uncertainty. With statistical thinking, it
allows us to accept and estimate levels of uncertainties (Steele, Andrews, &
Upton, 2012). In other words, statistics enables us to:
(b) Make inferences based on the data collected, which is another branch of
statistics called inferential statistics. Through inferential statistics, we can
confidently generalise a sample to the entire population (Howitt & Cramer,
2011). We can make predictions from the data collected using inferential
statistics.
The first example is a summary of data. The second example makes a prediction
of what will happen in the future.
SELF-CHECK 1.1
Statistics can be divided into two main branches as shown in Figure 1.2.
Now, see what Bulman (2013) says about the population and sample.
Population consists of all subjects (human or otherwise) that are being studied.
Sample is a group of subjects selected from a population.
SELF-CHECK 1.2
Looking at this example, the variable is the stress level, which has values from 0
to 10. The score is the value of your subordinateÊs answer. If your subordinate
answered 7, his score is 7; his score of 7 has a value of 7 on the variable called
„level of stress‰ (Aron et al., 2009).
Psychology research is about variables, values and scores (Aron et al., 2009),
which can be summarised in Table 1.1.
Score A particular personÊs value on the 0, 1, 2, 3, 4, 50, 89, 100, female, Catholic
variable
Apart from that, the survey questionnaire involves a rating of 0 to 10. The higher
the rating the more stress your subordinate experiences. This variable is known
as a numeric variable. As the name numeric states, it is a variable whose values
are numbers, which is also termed quantitative variable (Aron et al., 2009).
Quantitative variables can be further divided into two groups, discrete and
continuous (Bulman, 2013). Discrete variable is one that has a finite number or
countable number that cannot have values in between (Aron et al., 2009).
Examples of discrete variables are number of books you read, number of
employees in an organisation, number of cars you own, number of travels you
made last year and so on. On the other hand, continuous variables are defined as
variables with infinite number of values between any two values (Aron et al.,
2009). Common continuous variables are height, weight, time and temperature.
For instance, one may report his/her height as 169.3cm and weight 59.7kg.
However, due to the limitation of measuring device, the variables need to be
rounded up (such as 169cm) or put into a category (such as 168.5cm to 169.5cm)
during measuring (Bulman, 2013).
(i) Nominal variable is not a numeric variable, but its values stand for
names or categories (Aron et al., 2009).
(iii) For example, for the nominal variable gender, the values are female
and male. Another example is occupation, which has values, such as,
blue-collar worker, white-collar worker, professional, unemployed
and student.
(iv) Classifying cars according to the brand (e.g., Proton, Perodua, Toyota,
Honda, Mercedes Benz, Peugeot, etc.) is also an example of nominal
level of measurement (see Figure 1.3).
(v) Besides that, survey responses of „yes‰, „no‰, and „not sure‰, as well
as, resident zip codes which numbers have no meaningful order or
ranking, are also examples of nominal level measurement.
(i) Ordinal variable is one of the numeric variables in which the numbers
signify relative ranking (Aron et al., 2009). In other words, the scores
are placed in order and ranked from smallest to largest.
(ii) Data that are measured at this level can be arranged or placed into
categories with some order or ranks. However, the differences
between data values cannot be determined or are meaningless (Triola,
2010, p. 14). As stated by Bulman (2013), there is no precise
measurement of differences in this level of measurement.
(iii) One of the examples would be universityÊs results ranking (see Figure
1.4). Students are ranked with first, second upper, second lower and
third class honours. Students who are ranked with first class honours
do not know how many marks they were behind or ahead of other
students in the same rank or a different rank.
(ii) Data measured in this level are ranked data. There are precise
differences between data values.
(i) When the variable has an absolute 0 point that indicates complete
absence of the variable, the interval variable will be measured on a
ratio scale (Aron et al., 2009). Figure 1.6 shows the F1 Brazillian GP
First Practice Classification .
Ordinal Categories are ordered, but differences cannot be Exam grade, class
found or are meaningless. grade ranking, ranks
of universities
Looking at these different kinds of variables, researchers are to decide the types
of measurement that they want to use to measure their variables. „The level of
measurement selected affects the type of statistics that can be used with a
variable‰ (Aron et al., 2009, p. 5).
SELF-CHECK 1.3
The chart shows the number of vehicle occupants killed in the year 2013
in the United States.
Aron, A., Aron, E. N., & Coups, E. J. (2009). Statistics for psychology (5th ed.).
Upper Saddle River, NJ: Pearson.
Bluman, A. G. (2013). Elementary statistics: A brief version (6th ed.). New York,
NY: McGraw-Hill.
Howitt, D., & Cramer, D. (2011). Introduction to statistics in psychology (5th ed.).
Harlow, England: Pearson.
Schieman, S., Whitestone, Y. K., & Van Gundy, K. (2006). The nature of work
and the stress of higher status. Journal of Health and Social Behavior, 47(3),
242-257. doi:10.1177/002214650604700304
Steele, C., Andrews, H., & Upton, D. (2012). Statistics in psychology. Harlow,
England: Prentice Hall/Pearson.
INTRODUCTION
In the previous topic, you learned the operational definition of statistics, as well
as some basic terms used in statistics. As you read along, you may find yourself
knowing more about statistics than you think. Similar to what we always see in
articles, newspapers, websites and magazines, results are presented in tables,
graphs or pie charts.
After organising the data, the researcher must present it in a way that is easily
understood (Bluman, 2013). This can be done by creating statistical charts and
graphs. As stated by Bluman (2013), there are various types of charts and graphs
with different functions. In this topic, you will learn how to organise, summarise
and represent your data in a meaningful form.
ACTIVITY 2.1
In class or at work, record the mobile telephone brands that they have.
Present the data to your coursemates by using your own methods. After
that, compare your method with the methods that you learned
throughout this topic.
8 7 8 7
7 4 9 7
8 5 9 6
3 8 7 8
9 10 3 9
4 8 6 9
3 6 5 7
0 7 9 5
5 7 1 6
2 6 2 8
With the raw data, a researcher may organise it into a frequency distribution.
According to Bluman (2013), frequency distribution (or frequency table) is
defined as the organisation of raw data in table form, using categories (or classes)
and frequencies. The frequency of a class is the total amount of a specific value in
a particular category. In other words, a frequency table shows how many times
each class occurs (Aron, Aron, & Coups, 2009). An example of frequency
distribution for the above data set is shown in Table 2.2.
Stress
(a) Frequency (b) Stress Level Frequency
Level
0 1 0-2 Low 4
1 1 3-5 Moderate 8
2 2 6-8 High 21
3 3 9-10 Extremely High 7
4 1
5 4
6 5
7 9
8 7
9 6
10 1
With this frequency table, one may get some general information from the data.
For example, it can be said that the majority of working mothers in the study has
high levels of stress (or stress level of 7).
Looking at Table 2.2 (a), the classes in the distribution are 0, 1, 2, 3, ⁄ 10, whereas
the classes in Table 2.2 (b) are 0-2, 3-5, 6-8, and 9-10. These values are known as
class limits (Bluman, 2013). The data values 0, 1, 2 can be placed in the first class,
3, 4, 5 in the second class, and so on.
SELF-CHECK 2.1
Example 1
Stress Levels of Working Mothers
8 7 8 7
7 4 9 7
8 5 9 6
3 8 7 8
9 10 3 9
4 8 6 9
3 6 5 7
0 7 9 5
5 7 1 6
2 6 2 8
Step 1: From the collected data, list down each of the possible values, from
lowest to highest.
For example, referring to the Table 2.1 and 2.2, we assume that ratings
of the stress level range from 0-10. With this, the lowest value of the
stress level is 0, which is the lowest possible value, and 10 is the highest
possible value. All possible values need to be listed down even though
one of the values is not used. Any value that is not used can be
represented with a frequency of 0. For instance, if no one rated a stress
level of 3, value „3‰ still need to be included in the frequency table and
0 will be recorded as its frequency.
Step 2: Go through all the scores, draw a mark for each next to its value on the
list as showed in Figure 2.1.
Step 3: Draw a table showing how many times each value on the list is used to
get the frequency and sum up the numbers of marks beside the value as
demonstrated in Table 2.3.
Step 4: Calculate the percentage of scores for each value as indicated in Table
2.3.
To calculate the percentage, take the frequency for the value, divide it
by the total frequency and multiply by 100.
f
%= 100
n
f = frequency of the stress level
n = total numbers of values
3
%= 100 7.5%
40
Example 2
Distribution of Temperaments
Step 3: Mark the data and place the results in the „Tally‰ column.
Step 4: Total up the tallies and place the results in the „Frequency‰ column.
Step 5: Calculate the percentage of each class and place the results in the
„Percent‰ column.
5
Choleric % = 100 20%
25
Example 3
GraduatesÊ Ages in Applied Social Science Disciplines in OUM Year 2015
33 24 34 30 36 34 28 44 40 26
26 29 27 42 29 36 33 36 31 34
47 30 32 25 50 35 32 29 37 35
30 31 28 37 24 42 38 28 30 33
28 29 33 65 35 27 29 34 59 27
Find the range by subtracting the lowest value from the highest value.
R = 65 ă 24 = 41
Select the number of classes desired (usually between 5 and 20). In this
example, 8 is arbitrarily chosen.
R
Width =
number of classes
41
Width = 5.86
7
5.86 6.
Select a starting point for the lowest class limit. This can be the smallest
data value or any convenient number less than the smallest data value.
In this case, 24 is used. Add the width to the lowest score taken as the
starting point to get the lower limit of the next class. Keep adding until
there are 7 classes, as shown in Table 2.6.
Subtract one unit from the lower limit of the second class to get the
upper limit of the first class. Then add the width to each upper limit to
get all the upper limits.
30 ă 1 = 29
Find the class boundaries by subtracting 0.5 from each lower class limit
and adding 0.5 to each upper class limit as follows:
Step 3: Find the numerical frequencies from the tallies, as shown in Table 2.6.
n f 50
The frequency distribution shows that the class 29.5-35.5 contains the
largest number of graduates, followed by the class 23.5-29.5 with 17
graduates. Hence, most of the graduates (19) fall between 29.5 and 35.5
years old.
SELF-CHECK 2.2
(ii) Gender
(iii) Age
(iiii) Race
Cumulative frequencies are used to show how many data values are
accumulated up to and including a specific class. In Example 3, a total of 36
graduates are below 36 years old. 49 of the total graduates are 59 years old and
below.
Apart from the four frequency distributions mentioned earlier, there is also
relative frequency distribution. In this frequency distribution, relative frequency
(a proportion) or a percentage frequency (a percent) replaces the classÊs
frequency (Triola, 2010). The calculation of relative frequencies is demonstrated
in the Figure 2.2.
Table 2.8 shows the relative frequency distribution of number of cars in a family.
After the organisation of the raw data into a frequency distribution, one can then
analyse the data by looking for extreme values or what are call peaks. These
extreme scores show which class or classes have the most data values when
comparing the different classes (Bluman, 2013).
ACTIVITY 2.2
Create a table listing the watch brands that you can recall now. Then,
add in a column „do not know any of the brand‰ at the end of your
table. Subsequently, survey your coursemates to find out the watch
brands that each of your coursemates is aware of. Record the results
and construct a frequency table for your data.
SELF-CHECK 2.3
A categorical frequency distribution is used for data that can be classified into
specific categories, such as nominal- or ordinal-level data.
Aron, A., Aron, E. N., & Coups, E. J. (2009). Statistics for psychology (5th ed.).
Upper Saddle River, NJ: Pearson.
Bluman, A. G. (2013). Elementary statistics: A brief version (6th ed.). New York,
NY: McGraw-Hill.
INTRODUCTION
In the Topic 2, we had looked into the frequency distribution. It is the first data
organisation after data collection. Other than presenting them in table form, we
can also present the data in pictorial form, such as, charts, polygons, graphs and
so on. In this topic, you will learn pictorial presentation.
In statistics, graphs are used to illustrate and analyse the data (Bluman, 2013).
According to Bluman (2013), most lay people can understand the data showed in
pictorial form better compared to the data shown in tables. Graph presentations
are said to be more attractive to people as well. In addition, people also use
statistical graphs to note the trend of a situation or an issue over a duration,
discuss it and summarise it from the data presented (Bluman, 2013).
In the following, we will look at the most common graphs that are used in
research: histogram, frequency polygon, and cumulative frequency graph or
ogive (pronounced „o-jive‰). You may see the example for each graph in Figure
3.1(a) to Figure 3.1(c).
3.2 HISTOGRAMS
Histogram is defined as a graph made up of contiguous equal width vertical bars
(without gap). The horizontal scale indicates the classes of the data while the
vertical scale indicates the frequencies of the classes. The heights of the vertical
bars depend on the frequency values (Triola, 2010).
Step 1: Draw the x axis and y axis. Label the horizontal axis as x axis and
vertical axis as y axis.
Step 3: Write down the frequency values on the y axis and the class boundaries
on the x axis.
Step 4: Draw the vertical bars for each class by referring to the frequencies on
the y axis as the heights.
As the histogram in Figure 3.2 shows, the class with the greatest number of data
values (29) is 10.5-15.5 downloaded phone applications, followed by 23 for 15.5-
20.5 downloaded phone applications. The graph also has one peak with the data
clustering around it.
ACTIVITY 3.1
In class, each student record his or her time for holding their breaths.
Construct a frequency distribution and histogram for the times of both
males and females. Compare the results.
Step 1: Calculate the midpoints for each class by using the formula:
120 130
120 ă 130 125 2
2
130 ă 140 135 5
140 ă 150 145 16
150 ă 160 155 24
160 ă 170 165 31
170 ă 180 175 19
180 ă 190 185 3
Step 2: Draw the x and y axes. Write down the midpoint of each class.
Determine a suitable scale for the frequencies and then write down the
frequency values on the y axis.
Step 3: Mark the points by referring to the midpoints for the x values and the
frequencies as the y values.
Step 4: Use line segments to connect the adjoining points (see Figure 3.3). Start
the line from the beginning of the graph on x axis (i.e., the origin ă the
intersection of x axis and y axis), and end the line on x axis with the
same width between midpoints.
ACTIVITY 3.2
You have now just learned about histogram and frequency polygon.
Have you given thought of which of these graphs is better suited to
represent data or easier for people to understand the data? In statistics,
it is the researcherÊs choice to use what graph(s) to present their data
(Bluman, 2013). In class, discuss with your tutor and coursemates,
which graph is best to use for describing which types of data.
Step 2: Draw the x and y axes. Mark and label the x axis with the class
boundaries. Determine a suitable scale for the cumulative frequencies,
and then write it down on the y axis. In this example, a scale of 0, 10, 20,
30, 40, ⁄ will be used.
Step 3: Mark the cumulative frequency at each upper class boundary, as shown
in Figure 3.4. Upper boundaries are used since the cumulative
frequencies represent the number of data values accumulated up to the
upper boundary of each class.
Step 4: Use line segments to connect the adjoining points. Starting the line with
the lower boundary, 120, connect the adjoining points with line
segments, as shown in Figure 3.4.
As stated by Triola (2010), ogives are advantageous when you want to find out
the number of values that are below some particular values. For instance, if you
would like to find out how many students are shorter than 160cm, locate 160 on
the x axis, draw a vertical line up until it intersects the graph, and then draw a
horizontal line at that point to the y axis. The y axis value is 47 (Bluman, 2013), as
shown in Figure 3.5.
Table 3.4 summarises the steps for drawing histogram, frequency polygons, and
ogives.
Step 2 Choose a suitable scale for the frequencies or cumulative frequencies and
label it on the y axis.
Step 3 Represent the class boundaries for the histogram or ogive or the midpoint for
the frequency polygon, on the x axis.
Step 4 Plot the points and then draw the bars or lines.
Relative frequency graph is especially useful when there are large data sets or
when a researcher wants to see how often something happens (Andale, 2016). It
evaluates various classes against the whole. As stated by Bluman (2013), relative
frequency graph is used when the proportions of data values that fall into a given
class is more important than the actual number of data values that fall into that
class. For example, if a researcher wanted to compare the auto market production
in the world in 2014, the researcher may choose to use the relative frequency
graph. The reason is that there is a huge disparity between the countries in the
world, for instance, China produced 23,722,890 vehicles while Ukraine produced
28,751 vehicles (International Organization of Motor Vehicle Manufacturers,
2016).
Step 1: Calculate the relative frequency for each class by using the formula:
Class frequency
Frequency total
2
0 - 10.5 2 0.01
150
4
10.5 - 20.5 4 0.03
150
6
20.5 - 30.5 6 0.04
150
11
30.5 - 40.5 11 0.07
150
14
40.5 - 50.5 14 0.09
150
33
50.5 - 60.5 33 0.22
150
30
60.5 - 70.5 30 0.20
150
27
70.5 - 80.5 27 0.18
150
18
80.5 - 90.5 18 0.12
150
5
90.5 - 100.5 5s 0.03
150
150 0.99
Step 3: Draw the x and y axes. Mark and label the x axis with the class
boundaries. Determine a suitable scale for the relative frequencies and
then write it down on the y axis.
Step 4: For histogram, draw the vertical bars for each class by referring to the
relative frequencies on y axis as the heights, as shown in Figure 3.6.
Figure 3.6: Relative frequency histogram for kilometres driven per week
For frequency polygon, you will need to write down the midpoint of
each class before marking and connecting the points, as shown in Figure
3.7. The midpoints of each class are shown in Table 3.6.
Table 3.6: Midpoints of Each Class for Kilometre Driven per Week
Figure 3.7: Relative frequency polygon for kilometres driven per week
For cumulative relative frequency, add the frequency in each class to the total
frequency of the previous class. You may also find the cumulative frequencies
first and then convert each one to a relative frequency (see Table 3.7). Figure 3.8
shows the graph of cumulative relative frequency for kilometres driven per
week.
Figure 3.8: Cumulative relative frequency for kilometres driven per week
SELF-CHECK 3.1
A bar graph uses bars with equal width that may or may not be separated by
small gaps, to indicate the frequencies of the qualitative or categorical data
(Triola, 2010). Bar graphs can be drawn horizontally or vertically (Bluman, 2013).
Table 3.8: Average Money Spent during Chinese New Year (CNY)
RM
Gift for parents 2,000
CNY decorations and fireworks 1,500
Gifts for friends and relatives 500
Red packet (Ang Pao) money 1,000
Clothes shopping 1,300
Transportation fees 350
Reunion dinner 1,200
Step 1: Draw the x axis and y axis. Label the horizontal axis as x axis and
vertical axis as y axis.
Step 2: Decide your bar graph orientation. If you would like to draw a
horizontal bar graph, locate the frequency scale on the x axis. On the
other hand, if you would like to draw a vertical bar graph, then locate
the frequency scale on the y axis.
Step 3: Label the graph with details (for example, frequency values, classes).
Step 4: Draw the bars for each class by referring to the frequencies on x axis or
y axis (see Figure 3.9(a) & (b)).
Figure 3.9(a): Horizontal bar graph for money spent during Chinese New Year
Figure 3.9(b): Vertical bar graph for money spent during Chinese New Year
ACTIVITY 3.3
Recall your expenses for the previous month and construct a bar graph
for them.
Step 1: Order the data by placing the category with largest count first and
lowest at the bottom.
Step 2: Draw the axis x and axis y. Label the horizontal axis as x axis and
vertical axis as y axis.
Step 3: Draw the bars for each class by referring to the frequencies on y axis
(see Figure 3.10).
Table 3.10: Turnout Rates in Company Z for the Year 2008 to 2015
Step 1: Draw the x axis and y axis. Label the horizontal axis as x axis and
vertical axis as y axis.
Step 2: Label the x axis for years and the y axis for the number.
Step 4: Draw line segments to connect adjacent points. Do not attempt to fit a
smooth curve through the data points (see Figure 3.11).
Figure 3.11: Time series graph for turnout rates in company Z for the years 2008 to 2015
Time series graph is useful when a researcher wants to see the pattern or trend
for something. As Bluman (2013) stated, the researcher may observe the trend or
pattern of the time series graph, as an ascending line indicates an increase while a
descending line indicates a decrease over that period. In addition, the researcher
may also look at the steepness of the line, as a steeper line indicates faster
increase or decrease.
With time series graph, the researcher may also compare two or more sets of data
on the same graph (Bluman, 2013). This graph is known as compound time series
graph. As shown in Figure 3.12, the numbers of employees increased gradually
from 2008 to 2015 for both males and females.
Country Numbers
Australia 120,000
Brunei 90,000
USA 80,000
Britain 80,000
Canada 30,000
Singapore 460,000
Step 1: A circle has 360À. To find the proportional part of the circle for each
class, the frequency has to be converted by using the formula:
f
Degrees = 360À
n
Where f = frequency for each class and n = sum of the frequencies. The
total degrees should sum to 360À.*
120000
Australia 360À = 50.2À
860000
90000
Brunei 360À = 37.7À
860000
80000
USA 360À = 33.5À
860000
80000
Britain 360À = 33.5À
860000
30000
Canada 360À = 12.6À
860000
460000
Singapore 360À = 192.6À
860000
Total 360.1Àssss
f
%= 100
n
120000
Australia 100 = 13.9%
860000
90000
Brunei 100 = 10.5%
860000
80000
USA 100 = 9.3%
860000
80000
Britain 100 = 9.3%
860000
30000
Canada 100 = 3.5%
860000
460000
Singapore 100 = 53.5%
860000
Total 100%
Step 3: Next, use a protractor and a compass to draw the graph using the
appropriate degree measures found in Step 1, and label each section
with the name and percentages, as shown in Figure 3.13.
*Note: The degrees column does not always sum to 360Ĉ due to rounding.
†Note: The percent column does not always sum to 100% due to rounding.
SELF-CHECK 3.2
ACTIVITY 3.4
In your class, do a survey and record the numbers of siblings that each
of your coursemate has. After that, create a frequency table and
construct different types of graphs covered in this topic, that you think
is suitable to represent your data sets.
3.10 CONCLUSIONS
Graphs make data easier to be analysed and interpreted with visual aids. Lay
people can easily understand by looking at the numbers. Nevertheless, if one
drew a graph inappropriately, it will lead to a false conclusion (Bluman, 2013).
For example, a brand A skin product advertisement stated that 98% of the
customers showed improvement in their skin condition after using their product.
The advertisement then showed a graph similar to the one in Figure 3.14. The
graph shows the percentage of the customers who see improvement in their skin
condition after using their product and the percentage of its competitors. From
Figure 3.14, it seems to have a large difference. In fact, there is not much
difference.
If you pay attention to the graph, you will notice that the scale on axis y in Figure
3.14 has been cut off and starts at 90%. Now when you look at Figure 3.15 that is
redrawn using a scale that goes from 0 to 100%, the difference is hardly noticed.
In short, changing the units at the starting point on the y axis can convey a
different visual representation of the data (Bluman, 2013).
Figure 3.14: Graph of brand A improvement claim using a scale from 95 to 100%
In addition, the second most common types of bad graph are pictographs with
the drawings of objects (Triola, 2010). The reason is that the drawer can create
false impressions that mislead the reader. For example, Figure 3.16 shows the
number of people who own different pets. It seems that people own pet dogs
more than any other pets. This picture is misleading because the pictures are
drawn with different sizes. By looking at Figure 3.17 that is redrawn with same
size pictures aligned with one another, it is clear that most people own pet fish.
Frequency polygon is a graph that uses lines segments that connect the points
marked for the frequencies above the class midpoint values, to present the
data.
A bar graph uses bars with equal width that may or may not be separated by
small gaps, to indicate the frequencies of the qualitative or categorical data.
A Pareto chart is a kind of bar graph used for qualitative or categorical data.
It is advantageous when the researcher is trying to draw the readersÊ
attention to the categories that are more important.
When data are collected over an extended period of time, they can be
represented by a time series graph.
A pie graph is a circle that is divided into sections or wedges according to the
percentage of frequencies in each category of the distribution.
Andale. (2016). Relative frequency histogram: Definition and how to make one.
Retrieved from http://www.statisticshowto.com/relative-frequency-
histogram-2/
Bluman, A. G. (2013). Elementary statistics: A brief version (6th ed.). New York,
NY: McGraw-Hill.
Heymann, E. (2011). ASEAN auto market: Growing in the shadow of China and
India. Retrieved from www.dbresearch.com
Remarks
The steps in constructing graphs and tables are universal elements of the
statistics. They are written with reference to Bluman (2013).
INTRODUCTION
The previous topics revealed how you can organise raw information into
frequency distribution, and present the data by using a variety of pictorial
presentations or graphs. This topic shows the next step of statistical method ă
summarizing data. According to Bluman (2013), the most common statistical
methods to find the averages are called measures of central tendency. These
measures describe typical or representative scores. Although measures of central
tendency convey only part of the meaning of a set of scores or numbers, the
information regarding the center of distribution is important. The mean, median
and mode are used to reflect different notions of the center of a distribution. Each
can also be used over the other when attempting to give meaning to different
types of data. These measures are used in research, as well as, for describing data
to the public.
M=
X
N
M = symbol for mean; some people also use the symbol „ ‰ alternately
= symbol for „sum of‰, which means the total of all the numbers that
follow
= scores in the distribution of the variable X
X = sum of X, that is, the total of the scores in the distribution of the
variable X
N = number of scores in a distribution
(Aron et al., 2009)
Example 1
Suppose psychology students are participating in a research study to determine
whether time spent studying everyday results in higher grades. The number of
hours were as follows:
2, 4, 3, 2, 1, 0, 2, 3, 4, 1, 1, 0
2 + 4 + 3 + 2 + 1 + 0 + 2 + 3 + 4 + 1 + 1 + 0 = 23
23 12 = 1.92
The mean number of hours studied for this group of psychology students is 1.92.
In other words, on the average, each psychology student studies 1.92 hours per
day. Thus, the information for the 12 students is summarised by the single
number 1.92. The mean does not need to be a score in the distribution, but an
average of the scores. The mean is viewed as the balance point of the scores
(Aron et al., 2009).
Next, what about grouped data? How do we calculate the mean for grouped
data? Let us look at Table 4.1 of Example 2.
Example 2
Table 4.1: GraduatesÊ Ages in Applied Social Science Disciplines in OUM Year 2015
23.5 29.5
23.5-29.5 17 Xm = = 26.5
2
29.5 35.5
29.5-35.5 19 Xm = = 32.5
2
35.5 41.5
35.5-41.5 7 Xm = = 38.5
2
41.5 47.5
41.5-47.5 4 Xm = = 44.5
2
47.5 53.5
47.5-53.5 1 Xm = = 50.5
2
53.5 59.5
53.5-59.5 1 Xm = = 56.5
2
59.5 65.5
59.5-65.5 1 Xm = = 62.5
2
n = f = 50
M=
X =
1685
= 33.7 years old
N 50
In order to find the mean for grouped data, we assume that the mean of all the
raw data values in each class is the same as the midpoint of the class (Bluman,
2013). In fact, this is not accurate as the average of the raw data values in each
class usually will not be exactly equal to the midpoint. Nevertheless, we may use
this procedure to obtain an acceptable estimate of the mean because some values
fall above the midpoint and other values fall below the midpoint for each class,
and thus, the midpoint represents an estimate of all values in the class (Bluman,
2013).
SELF-CHECK 4.1
There are times when the median is better than the mean as a representative
value for a group of scores. This occurs when there are a few extreme scores that
would strongly influence the mean but would not affect the median (Aron et al.,
2006).
Example 1
Find the median for studentsÊ grades in a psychology paper. The scores are 75,
47, 55, 63, 68, 89, 85, 80, 60, 58, 62, 78, 70, 69, 63, 64 and 74.
47, 55, 58, 60, 62, 63, 63, 64, 68, 69, 70, 75, 78, 78, 80, 85, 89
Ę
Step 2: Select the middle score: 68
Example 2 is to determine the median when you have an even number of scores.
Example 2
Imagine six psychology students who are asked to record how many hours of
sleep they get in one week, and if less sleep results in higher grades, as part of a
research study. Hours recorded are 28, 50, 56, 50, 49 and 50.
Step 2: Find out how many scores there are to the middle score by adding 1 to
the number of scores and dividing by 2.
Middle score = 6 2 = 3
Step 3: Count up to the middle score or scores. If you have one middle score,
this is the median. If you have two middle scores, the median is the
mean, or average of these two scores.
50 50
MD = = 50
2
The third and fourth scores are both 50s. Hence, in either way, the
median will be 50.
Unlike the mean, the median will be less influenced by the extreme score of 28,
thus giving you a more representative value of the scores.
The Mode
The mode represents another measure of central tendency. The mode is the most
frequently occuring value in a distribution (Aron et al., 2006). One way to think
of the mode is that it is the value with the largest frequency in a frequency table.
Please remember that the mode and the mean can be different and thus the
frequency is not perfectly symmetrical.
The mode is very useful when studying nominal variables, when the mean and
the median are not useful. For example, how many students major in business or
psychology at a given university. With mode, you may see the studentsÊ
preference in studying at a given university.
Example 1
Find the mode of the number of books students read in two months. The number
of books are
2, 2, 1, 4, 3, 0, 1, 2, 1, 0, 2
Tips: Arranging the data from lowest to highest may help you to identify the
mode easier.
0, 0, 1, 1, 1, 2, 2, 2, 2, 3, 4
When the data are grouped, the mode for it is known as the modal class. Modal
class refers to the class with the largest frequency (Bluman, 2013).
Example 2
Table 4.2 shows kilometres driven per week.
The modal class is 50.5-60.5 as the frequency value is the largest. According to
Bluman (2013), the midpoint of the class is used sometimes, rather than the class
boundaries. Thus, the mode could also be given as 55.5km per week.
SELF-CHECK 4.2
1. Define the mode. What type of data is most appropriate to use the
mode?
2. Figure the mean, mode and median for the following data:
46, 52, 55, 32, 44, 48, 61, 59, 44, 49, 44
ACTIVITY 4.1
Find the mean, median and mode for the examples given as a part of
this activity.
4.5 ORGANISATION
When data is arranged in ascending or descending order, it can be divided into
different parts called deciles, percentiles and quartiles. This extension of the
median formula divides data into equal parts. This division, which helps give
clearer meanings to the numbers, also highlights the relationship between the
divided parts.
4.6 PERCENTILES
Percentiles are measures that divide a group of data into 100 parts. These are
used when assessing performance and comparing the results between different
variables (Glass& Hopkins, 1996).
Example 1
A 20-point quiz was given to 20 students. The scores are as shown below. Find
the percentile rank of a score of 14.
18, 15, 10, 19, 12, 20, 17, 15, 14, 10, 17, 18, 16, 12, 13, 15, 17, 12, 17, 11.
10, 10, 11, 12, 12, 12, 13, 14, 15, 15, 15, 16, 17, 17, 17, 17, 18, 18, 19, 20
11 0.5
Percentiles = 100 57.5 percentile
20
Therefore, a student who obtained a score of 14 did better than 57.5% of the class.
Example 2
Using the scores in Example 1, find the value corresponding to the 35th
percentile.
10, 10, 11, 12, 12, 12, 13, 14, 15, 15, 15, 16, 17, 17, 17, 17, 18, 18, 19, 20
20 35
Step 3: c = 7
100
Step 4: Begin at the lowest score, count up to the seventh value, which is 13. If c
is not a whole number, round it up to the next whole number. For
example, if c is 6.5, you may round it up to 7.
4.7 QUARTILES
The values that divide the data set into four equal parts after it has been arranged
in ascending order are called quartiles, which can be separated by Q1, Q2, Q3. In
other words, Q1 can be marked as 25th percentile, Q2 can be labelled as 50th
percentile, which is also the median and Q3 corresponds to 75th percentile, which
are shown below:
8 10
MD / Q2 = =9
2
Step 3: Find the median of the data with values less than 9.
3, 5, 7, 8
57
Q1 = =6
2
Step 4: Find the median of the data with values greater than 9.
10, 12, 22, 27
12 22
Q3 = = 17
2
4.8 DECILES
This is when variables are divided into ten equal parts. For example, 1/10 would
represent a way of describing a sample of scores to give meaning for a particular
population. Calculation of deciles are similar to quartiles above. The difference is
that deciles is divided into 10 groups, but quartiles is divided into 4 groups.
In a nutshell, taken all together, the relationships among percentiles, deciles and
quartiles are shown below (Bluman, 2013):
(a) Deciles are represented by D1, D2, D3, ⁄ , D9 and they denoted as P10, P20,
P30, ⁄ , P90.
(b) Quartiles are represented by Q1, Q2, Q3 and denoted as P25, P50, P75.
SELF-CHECK 4.3
2. Figure the mean, mode, and median for the following data:
46, 52, 55, 32, 44, 48, 61, 59, 44, 49, 44
The mean is the average score, which is the most commonly used measure of
central tendency.
The median is the middle score ă the score at the 50th percentile.
Percentiles are measures that divide a group of data into 100 parts.
Quartiles are the values that divide the data set into four equal parts after it
has been arranged in ascending order.
Aron, A., Aron, E., & Coups, E. (2006). Statistics for psychology (4th ed). Upper
Saddle River, New Jersey: Pearson Education.
Aron, A., Aron, E. N., & Coups, E. J. (2009). Statistics for psychology (5th ed.).
Upper Saddle River, NJ: Pearson.
Bluman, A. G. (2013). Elementary statistics: A brief version (6th ed.). New York,
NY: McGraw-Hill.
Glass, G.V. & Hopkins, K.D. (1996). Statistical Methods in Education and
Psychology, (3rd ed.). Boston: Allyn & Bacon.
INTRODUCTION
In the previous topic, we have discussed measures of central tendency. With
central tendency, we get to identify the central position of the data set that
enables us to see the general ideas of the data. However, measures of central
tendency (such as, mean, mode and median) do not describe the differences
among scores. Other statistical measures are needed to do this, which are known
as measures of variability or dispersion. In other words, the variability describes
the extent to which scores in a distribution differ from one another (Heiman,
2006). We will discuss the range, inter quartile range, variance and standard
deviation as a way of explaining distinctions between populations in this topic.
ACTIVITY 5.1
5.1 VARIABILITY
Variability refers to the distribution around the measure of central tendency. For
example, you could ask for the average age of learners in this statistics class. Let
us say that the average age is 35. Nevertheless, it does not mean that everyone in
the class is 35 years old. Instead, some could be younger, say 23 years old, and
some could be older, say 38 years old. This gives you information about the
spread of ages around the mean age of 35 and gives you more information
regarding the population of the class. More specifically, the variability describes
people who are not at the mean age of 35.
Interquartile range = Q3 ă Q1
For instance, you have a set of data: 63, 97, 87, 20, 256, 58, 33, 152, 179, 138. In
order to find interquartile range, you will need to:
20, 33, 58, 63, 87, 97, 138, 152, 179, 256
Step 2: Find Q1
Q3 ă Q1 = 152 ă 58 = 94
SELF-CHECK 5.1
1. Explain why the range might not be the best estimate of variability.
2. Find the range and interquartile range for the following data:
88, 35, 290, 173, 30, 49, 85, 271, 295, 220, 76, 64, 59, 115, 234
5.4 VARIANCE
The variance explains how spread out the scores is around the mean. To be
precise, the variance is the average of each scoreÊs squared difference from the
mean.
X M
2
SD 2
N
Brand X Brand Y
6 17
30 25
25 15
16 17
20 20
11 14
M
X
6 30 25 16 20 11
18
N 6
Step 2: Subtract the mean from each data value. This gives each scoreÊs
deviation score. The deviation score is how far away the score is from
the mean.
Brand X
6 ă 18 = ă 12
30 ă 18 = +12
25 ă 18 =+7
16 ă 18 =ă2
20 ă 18 =+2
11 ă 18 =ă7
Brand X
( ă 12)2 = 144
(+ 12) 2 = 144
(+ 7) 2 = 49
(ă 2) 2 =4
(+ 2) 2 = 49
(ă 7) 2 =4
Step 4: This gives each scoreÊs squared deviation score. Add up the squared
deviation scores. This total is called the sum of squared deviations.
This gives the average (the mean) of the squared deviations, called the
variance.This is an important relationship. The square root of the variance results
in the standard deviation. If the process is reversed and the standard deviation is
squared, you will get the variance. Also, the closer to the mean, the smaller will
be the variance, which then indicates that farther away from the mean, the
variance will be larger.
Copyright © Open University Malaysia (OUM)
TOPIC 5 MEASURES OF DISPERSION OR VARIABILITY 81
ACTIVITY 5.2
Using the previous example, find the variance for Brand Y. Compared
to Brand X, which brand do you think is better? State your reason.
SD SD 2
Step 2: Take the square root. The standard deviation is the positive square root
of the variance.
SELF-CHECK 5.2
Number of clients: 18, 30, 14, 13, 20, 35, 12, 10, 21, 38, 15, 15, 19
M
X
18 30 14 13 20 35 12 10 21 38 15 15 19
N 13
260
20
13
Squared
Number of Deviation
îă Mean = Deviation
Clients Scores
Score
18 îă 20 -2 4
30 îă 20 10 100
14 îă 20 -6 36
13 îă 20 -7 49
20 îă 20 0 0
35 îă 20 15 225
12 îă 20 -8 64
10 îă 20 -10 100
21 îă 20 1 1
38 îă 20 18 324
15 îă 20 -5 25
15 îă 20 -5 25
19 îă 20 -1 1
¡=0 954
X M
2
954
Variance SD 2
73.38
N 13
SD SD 2 73.38 8.57
This result tells us that the number of clients a psychologist has varies
from the mean of 20 by an average of 8.57.
Here are the steps to find variance and standard deviation of grouped data.
11-15 1 13
16-20 4 18
21-25 6 23
25-30 7 28
31-35 5 33
35-40 2 38
Midpoint
Class Limits Frequency (f) f X m f X m2
(Xm)
11-15 1 13 13 169
16-20 4 18 72 5,184
21-25 6 23 138 19,044
25-30 7 28 196 38,416
31-35 5 33 165 27,225
35-40 2 38 76 5,776
25 660 95,814
Step 4: Total up the frequency, and the scores you found for steps 3 & 4.
Sum of frequency = n
Sum of f X m f X m
Sum of f X 2m f X 2m
n f X 2m f X m
2
SD
2
n ( n 1)
Step 6: Take the square root of variance to get the standard deviation.
SD SD 2 3266.25 57.15
(i) The larger the variance and standard deviation, the more varied the
data. This information is helpful when comparing two or more data
sets to determine which is more variable.
(ii) For example, looking at the example in Subtopic 5.4, you will find
the standard deviation of Brand X is larger than Brand Y. With this,
you may say that Brand Y is better. This is because, when the means
are equal, the larger the variance or standard deviation is, the
greater the variation of the data. Simply put, the differences in
Brand Y is smaller, which further suggests that Brand Y is more
stable.
(c) Determining the Number of Data Values that Fall within a Specified
Interval in a Distribution
As stated by Bluman (2013), ChebyshevÊs Theorem shows that at least
75% of the data values will fall within 2 standard deviations of the mean.
ACTIVITY 5.3
If you had a coin and decided to flip it 5 times in as a group and 6 times
in another group, how would one determine the standard deviation?
Record the outcomes of both Group A and Group B, find the variance,
and the standard deviation. Then explain why this method might seem
more difficult.
5.9 SKEWNESS
A description of a distribution should include its central tendency, variability,
and also the degree of asymmetry. This lack of equivalence between aspects of
the data is referred to as skewness. Skewness can give visual meaning to data
that can sometimes be imprecise through numbers or words in isolation.
Skewness can influence measures of central tendency. For example, when
distributions are positively skewed, the mean will have the largest value and the
mode the lowest. This is reversed when distributions have negative skewness.
Figure 5.1 illustrates this, which will be discussed in more detail in later topics.
The variability of scores within a set can be described by using variance and
standard deviation.
The purpose of variability is to determine how spread out the scores are in a
distribution.
There are four measures of variability, which are the range, the interquartile
range, the variance and the standard deviation.
The range is the difference between the largest and smallest score in a
distribution.
The interquartile range is the difference between the third quartile (Q3) and
the first quartile (Q1).
Variance and standard deviation are the most frequently used measures of
variability.
Deviation Variance
Range Standard Deviation
Interquartile Range Skewness
Variability
Bluman, A. G. (2013). Elementary statistics: A brief version (6th ed.). New York,
NY: McGraw-Hill.
Glass, G.V. & Hopkins, K.D. (1996). Statistical methods in education and
psychology (3rd ed.). Boston: Allyn & Bacon.
Gravetter, E. J., & Wallnau, L. B. (2009). Statistics for behavioral sciences (8th ed.).
Belmont, CA: Wadsworth, Cengage Learning.
Heiman, G. W. (2006). Basic statistics for the behavioral science (5th ed.). Boston,
MA: Houghton Mifflin Company.
INTRODUCTION
The purpose of psychological research involves the examination of effectiveness
of a particular procedure related to a theoretical construct. However, research is
not designed to give a total truth, but to explain if a certain phenomenon occurs
more or less likely. Total certainty does not exist and this is a key element of how
statistics is used in psychology. Understanding normal distribution, often
referred to as the normal curve, and how this distribution occurs naturally, and
its comparison to research results allows for statistical meaning to be given to
various aspects of research results. How to determine the potential for an
outcome and why entire populations cannot be used in psychological research
One question regarding the normal curve relates to its occurrence in nature. The
most basic answer is yes and this type of distribution occurs frequently in
various settings. How was this noticed? Late in the nineteenth century, Francis
Galton took systematic measurements of many physical, psychological, and
psychomotor characteristics on large samples of people. What he found was that
his findings were very close to a normal distribution. His work soon began to
influence other peopleÊs research and it was found that the findings of other
types of research were very close to a normal distribution. For example, the daily
high temperatures for a particular date in the past 100 years will follow a normal
distribution. Another example could be the reaction times of 12 year old children
to a certain stimuli; again, the results would closely follow a normal curve.
SELF-CHECK 6.1
6.2 PROBABILITY
Probability is used to test a statistical inferential statement. In psychology, this
probability might be very important as a psychologist uses probability to reach a
conclusion in a research (Aron et al., 2009).
Suppose that you say there is a 95% chance that it will rain tomorrow. You could
be using a kind of relative frequency interpretation. This would imply that if you
were to check the weather on that particular day in the past you would find it
would rain 95% of the time. However, what you mean is probably more
subjective: on a scale of 0% to 100%, you would rate your confidence or it raining
at 95%. Alternatively, you would feel that it will rain and rate it at 95%.
One example is when you have a six sided dice; the probability of getting 1 is
1/6, or 0.17. A more complex example could involve a statistics class that has 200
students, and 30 of them are male. If you were to pick somebody from the class
randomly, what is the probability of picking a male student?
Step 3: Divide the number of possible successful outcomes by the number of all
possible outcomes.
30
Probability = = 0.15
200
SELF-CHECK 6.2
2. If you had two, six sided dice, what is the probability of getting
two ones, often called snake eye, when rolled four times?
For example, if you wanted to know the rate of depression among rural
Malaysians, one cannot interview every person in a rural setting in Malaysia. The
strategy in almost all psychology research is to study a sample of individuals
who are believed to be representative of the general population. Realistically,
researchers try to study people who do not differ from the general population in
any systematic way that should matter for that topic of research. This allows for a
more manageable way of constructing and then conducting psychological
research.
ACTIVITY 6.1
(a) How many students would you survey? Explain your sample
numbers and sampling method(s).
Nonetheless, there has been a persistent line of criticism of whether nature really
packages itself so neatly and normally. Various researchers in the USA have
studied the results of student tests, which help in university placement, and
found the distribution did not follow a normal curve, no matter how large or
small the sample size. However, these results did not provide answers to why a
normal distribution did not occur, thus an answer that simply states: it is not
known. Although the vast majority of statistical analysis does result in normal or
very close to normal distribution, one can only make the assumption that in
psychology, variation occurs because of the nature of variation in how one
interacts with its environment.
SELF-CHECK 6.3
Aron, A., Aron, E. N., & Coups, E. J. (2009). Statistics for psychology (5th ed.).
Upper Saddle River, NJ: Pearson.
LEARNING OUTCOMES
By the end of this topic, you should be able to:
1. Identify the steps of hypothesis testing;
2. State null hypothesis, research hypothesis, level of significance, test
statistic, p value, and statistical significance;
3. Differentiate t-test for a single sample, and t-test for dependent means;
4. Perform appropriate t-tests depending on the research;
5. Compute the t-tests;
6. Define effect size and statistical power; and
7. Report statistics in research articles.
INTRODUCTION
In this topic, you will learn hypothesis testing procedures, particularly with an
unknown population variance ă t-test. As stated by Creech (2016), t-test may be
the most common test used to test the hypothesis.
To begin, let us get some ideas on hypothesis testing. According to Aron et al.
(2009), hypothesis testing refers to the procedure used to determine if the results
of the research is „statistically significant‰, which means having support for a
particular theory or practical innovation. In other words, the aim of hypothesis
testing is to examine if the obtained difference between the data and the
hypothesis is significantly greater than would be expected by chance (Gravetter
& Wallnau, 2009).
Often, when researchers conduct a study, they want to learn about an unknown
population; they usually compare two or more groups of scores with one
another, without any direct information on the populations (Aron et al., 2009). In
this situation, the researcher may use t-test instead of z-test (which you will learn
in Topic 9 later on the statistical procedures that allow researchers to test
hypotheses with a known population), by comparing the mean of a single sample
to a population with a known mean but an unknown variance (Aron et al., 2009).
As you read along, you will find that t-test and z-test are very similar.
For instance, a manager of a flower business wants to find out if his companyÊs
delivery service is compatible to the competitorÊs who advertises delivery time of
30 minutes. Hence, the manager gathers a random sample of the companyÊs
flowers delivery times, and then uses one sample t-test to examine if its mean
delivery time is significantly lower than its competitors.
One sample t-test operates similar to a z-test (Aron et al., 2009). The only
difference between a t-test and a z-test is that the z-test uses the actual standard
deviation () or the population variance (2), whereas the t-test uses the
corresponding sample standard deviation or variance, as the population value is
unknown (Gravetter & Wallnau, 2009). In other words, z -test requires you to
know the values for both mean and standard deviation (or variance) of the
population, in order to compare the sample mean and calculate the standard
error (Gravetter &Wallnau, 2009).
a great deal of variation and vice versa. Therefore, it is said to be possible to use
the sample scores variation (or standard deviation) to infer the population scores.
Degrees of freedom = df = n ă 1
X M
2
ss
S2
N 1 df
For example, imagine that you want to study whether students skip more classes
when they stay in hostel compared to those who stay with family. Suppose you
have the informal survey from your collegeÊs management which shows that the
students skipped an average of 2 classes per 14 week semester. You then
randomly pick 20 students who stay in hostel and ask how many classes they
skipped in a semester (or check their attendance record). Your result is that
students who stay in hostel skipped an average of 4 classes per semester.
X M
2
Step 2: In order to get the estimated population variance, you will need to
divide this sum of squared deviation scores by degrees of freedom.
ss 872
S2 = = = 45.89
df 20 1
SELF-CHECK 7.1
S2
Variance of the distribution of means: S M2
N
Looking back at the previous example, the sample size was 20 and the estimated
population variance was 45.89 as computed. What is the variance and standard
deviation of the distribution of means? Let us see the calculation.
S 2 45.89
(a) Variance of the distribution of means: S M2 2.29.
N 20
Be aware that the symbols for the variance and standard deviation of the
distribution of means use S instead of when it is with an estimated population
variance (Aron et al., 2009). The symbols of different types of standard deviations
and variances are shown in Table 7.1.
Figure 7.2: Cut-off scores for t distributions with 1 through 19 degrees of freedom
If your one-tailed test is testing whether the mean of Population 1 is more than
the mean of Population 2, the cut-off t score is positive, and vice versa (Aron et
al., 2009). Looking at our example on students who skipped classes, assume that
we are testing whether students staying in hostels (Population 1) skip more
classes than students who stay with family (Population 2). This means our cut-off
t score is positive. Therefore, you will reject the null hypothesis if your sample
mean is 1.729 or more standard deviations above the mean on the comparison
distribution.
In addition, as you can see in the t table in the Appendix A, the degrees of
freedom range from1 through 30, then 35, 40, 45, 50, 55, and so on. What if your
study has degrees of freedom of 53? As stated by Aron et al. (2009), it is safer to
use the nearest degrees of freedom to yours on the table with those less than
yours. In this case, a study with 53 degrees of freedom, you would use the cut-off
for df = 50.
M 42
t 1.32
SM 1.51
In our example, looking at the formula above, the t score would be 1.32. It is
obtained by subtracting the population mean from the sampleÊs mean and then
dividing by the standard deviation of the distribution of means based on an
estimated population variance.
In our example, the cut-off t score was 1.729 and the actual t score for the sample
was 1.32. Since the calculated t value is lower than the t value from the table, the
null hypothesis cannot be rejected, and this study is said to be inconclusive.
Aron et al., (2009) has stated the steps of hypothesis testing as follows:
Research hypothesis:
Population 1 produces fewer screws than population 2.
Null hypothesis:
Population 1 will not produce fewer screws than population 2.
M = (20 + 22 + 22 + 20) 4 = 21
X M
2
ss
S 2
N 1 df
X M
2
= (20 ă 21)2 + (22 ă 21)2 + (22 ă 21)2 + (20 ă 21)2
=4
X M
2
4
S2 1.33
N 1 4 1
S 2 1.33
SM2
0.33
N 4
S M S M2 0.33 0.57
M 21 23
t 3.51
SM 0.57
The t of -3.51 is more extreme than the needed t of 2.343. Hence, the null
hypothesis is rejected and research hypothesis is supported.
SELF-CHECK 7.2
The average amount of milk collected per week is 126kg. Test the
hypothesis that a sheep produces more than 120kg milk per week by
demonstrating the hypothesis testing steps.
ACTIVITY 7.1
Count the contents of each item, and examine the hypothesis that your
sample mean differs significantly from the value stated.
In this situation, we cannot use the one sample t-test but to use t-test for
dependent means. The procedure of obtaining t-test for dependent means is
similar to the procedure in obtaining one sample t-test. However according to
Aron et al., (2009), t-test for dependent means:
(a) Use difference score (D value), which is the difference of two values,
instead of raw scores; and
(b) Assume that the mean difference for the general population is zero.
Difference score = D = X2 ă X1
Research hypothesis:
Social skills improved after the social skills training program.
Null hypothesis:
No improvement in social skills after the social skills training program.
108
M= = 18
6
ss 596
S2 119.2
df 5
S 2 119.2
SM2
19.87
N 6
S M S M2 19.87 4.46
M 18 0
t 4.04
SM 4.46
The t of 4.04 is more extreme than the needed t of 2.015. Hence, the null
hypothesis is rejected and research hypothesis is supported.
ACTIVITY 7.2
Review and compare z test, t-test for a single sample and t-test for
dependent means with your tutor and coursemates. Create a
comparison table for these three tests.
Nevertheless, Gravetter and Wallnau (2009) said that this normality assumption
is least important especially when there is a large sample. In other words, a
normal population distribution is important when the sample size is small. To be
safe, one just needs to use a larger sample. In addition, Aron et al. (2009) argued
that we have no idea if the population is normally distributed when we do
research. It is almost impossible to have equal variances for two populations.
Fortunately, statisticians found that distributions in psychology research are
always close to a normal curve and that one can still obtain accurate results with
t-tests even when the population is far from normal (Aron et al., 2009).
SELF-CHECK 7.3
1. Describe some situations which are suitable for using t-test for
independent means.
Let us say the first groupÊs mean score is 110 and the second groupÊs mean score
is 90. We know that there is a difference of 20. However, how big is this
difference? The idea of effect size is that if there are no overlaps and all students
in the first group obtain a higher score than those in the second group, the
difference is very significant, and vice versa (Coe, 2002).
1 2
d
d is the symbol of effect size, also known as CohenÊs d.
is the population mean
is the standard deviation
Looking at the experiment of sleep and learning, suppose the standard deviation
is 24, the effect size would be:
1 2 110 90
d 0.83
24
Table 7.3: Summary of CohenÊs Effect Size Conventions for Mean Differences
In terms of the effect size conventions, the sleep and learning experiment has a
large effect size. In other words, it says that the effect size of the experiment is
very substantial.
(i) The bigger the effect size, the greater the power there is in the study.
In other words, if there is a big mean difference in the population, the
chances of getting a significant result in the study is higher; and
Table 7.4: Approximate Power for Studies Using the t-test for Dependent
Means in Testing Hypotheses at the 0.05 Significance Level
Effect Size
Difference Scores in
Small Medium Large
Sample (N)
(d = 0.20) (d = 0.50) (d = 0.80)
One-tailed test
10 0.15 0.46 0.78
20 0.22 0.71 0.96
30 0.29 0.86 *
40 0.35 0.93 *
50 0.40 0.97 *
100 0.63 * *
Two-tailed test
10 0.09 0.32 0.66
20 0.14 0.59 0.93
30 0.19 0.77 0.99
40 0.24 0.88 *
50 0.29 0.94 *
100 0.55 * *
*Power is nearly 1
Suppose a researcher expects a medium effect size. He uses the significance level
of 0.05, two-tailed, and with 30 participants, the study would have a power of
0.77. In other words, it means that there is a 77% chance that this study will come
out significant if the research hypothesis is true, with medium effect size.
Table 7.5: Approximate Number of Research Participants Needed for 80% Power for the
t-test for Dependent Means in Testing Hypotheses at the 0.05 Significance Level
Effect Size
Small Medium Large
(d = 0.20) (d = 0.50) (d = 0.80)
One-tailed 156 26 12
Two-tailed 196 33 14
Imagine that you would like to conduct a research, in which you expect a
medium effect size and use the significance level of 0.05, one-tailed. By referring
to the table, you would need 26 participants to obtain 80% power.
SELF-CHECK 7.4
Suppose you want to carry out a study to predict the mean of the
population of difference scores to be 30 and the population standard
deviation 60. You plan to test significance using a t-test for dependent
means, one-tailed, with a significance level of 0.05.
Testing the same group of participants twice often yields high power in a study,
compared to a study that separates the participants into two groups to be tested
separately (Aron et al., 2009). The reason of having higher power in repeated-
measure design is that the standard deviation of difference scores is typically
low. This leads to a larger effect size and increases the power (Aron et al., 2009).
In contrast, Cook and Campbell (1979, as cited in Aron et al., 2009) expressed that
testing the same group of people without any control group is a weak design
because it may have variations in factors that will influence the results. Some
examples that may affect the results in repeated-measures design are matured or
improved participants, participants drop-out, order effects, and so on.
In addition, Bates College (2011) stated that the mean is usually reported together
with a measure of variability (such as standard deviation or standard error of the
mean). Nevertheless, standard deviation of difference scores is seldom reported
in articles (Aron et al., 2009). Other than that, researchers may also report the
direction of the relationship (such as, one-tailed or two tailed) in their journal
articles. According to Aron and his colleagues (2009), if the direction is not
specified, we will assume the research as two-tailed test. To give an example, a
researcher may report the results like „The sample of employees (M = 18, SD =
2.76) was significantly different from the known mean of 11 for the employees as
a whole, t(24) = 2.79, p < 0.05, one-tailed.‰
For t-test for dependent means, a researcher may report the results as „There was
a significant growth in social skills level, increasing from a mean of 12.9 before
participating in social skills training program to a mean of 15.2 after attending
the social skills training program, t(9) = 5.2, p < 0.05‰.
SELF-CHECK 7.5
When the population variance is unknown, use a formula that divides the
sum of squared deviation scores by the degrees of freedom (df = N ă 1) to
estimate the population variance from the scores in the sample.
t-test for dependent means is used when you have two scores for each
participant or when you have scores from pairs of participants. Hypothesis
testing procedure is the same as t-test, but to look for a difference or change
in score for each participant.
The effect size of a study using a t-test for dependent means is the mean of
the difference scores divided by the standard deviation of the difference
scores.
Statistical power of a research is the probability that the study will produce a
statistically significant result if the research hypothesis is true.
Aron, A., Aron, E. N., & Coups, E. J. (2009). Statistics for psychology (5th ed.).
Upper Saddle River, NJ: Pearson.
Coe, R. (September 25, 2002). ItÊs the effect size, stupid: What effect size is and
why it is important. Retrieved from http://www.leeds.ac.uk/educol/
documents/00002182.htm
Gravetter, E. J., & Wallnau, L. B. (2009). Statistics for behavioral sciences (8th ed.).
Belmont, CA: Wadsworth, Cengage Learning.
APPENDIX A
INTRODUCTION
In Topic 7, we learned t-test for independent means, which compares two groups
of scores from two separate groups of people. In this topic, we are going to look
at the association between two groups of scores correlation. In other words,
correlation, which is a statistical procedure, allows us to measure and describe a
relationship between two variables (Aron et al., 2009).
ACTIVITY 8.1
Spurious Correlation
Using your imagination, together with your conceptual idea on
correlation, think of two humorous variables that can be linked together.
Share in the forum or with your coursemates.
http://mnstats.morris.umn.edu/introstat/public/instruction/causatio
n.html
Step 1: Draw the x axis and y axis. Determine which variable to put on which
axis.
Step 2: Determine suitable range of values to use for each variable and mark
them on the axes.
(a) Numbers are to start from where the axes meet, with a value of 0
and continue to the highest value in the measure.
(a) Mark a dot by referring to the value on the x axis and the value on
y axis for each pair of scores.
(b) When the same pair of scores occurs more than one time, put a
second dot as near as possible to the first dot. The dot may be
touching the first dot, but you need to make it clear that there are
two dots in the same place. Otherwise, put number 2 in that place.
SELF-CHECK 8.1
Figure 8.2: Scatterplot from Figure 8.1 with line drawn to show general trend
On the other hand, when one variable increases but the other variable decreases,
it is said to have a negative linear relationship (Gravetter& Wallnau, 2009). It is
represented with a straight line going down a graph, as shown in Figure 8.3.
Figure 8.3: Negative relationship between hot chocolate sales and temperature
Nonetheless, the general pattern between the two variables sometimes does not
follow a straight line but a more complex pattern, where one variable increases
and the other variable increases and decreases at the same time. This is known as
a curvilinear relationship (Aron et al., 2009). Figure 8.4 shows an example of a
curvilinear relationship graph.
8.2.2 No Correlation
So far, we have looked at positive, negative, and complex relationships between
two variables. We have to be aware that there is also the possibility of having no
correlation between two variables. As shown in Figure 8.5, there is no line, but
dots are distributed everywhere.
(a) Strong/large linear correlation: When the dots fall close to a straight line;
(b) Weak/small correlation: When the dots fall far from a straight line or when
you can hardly tell that there is correlation;
(d) Perfect linear correlation: When all the dots fall exactly on the straight line.
ACTIVITY 8.2
Covariability of X and Y
Variability of X and Y separately
SP X M x Y M y
On the other hand, SP can also be calculated with the computational formula:
SP XY
X Y
n
Mx = (2 + 4 + 8 + 10) 4 = 6
My = (6 + 12 + 8 + 14) 4 = 10
X Y XY
2 6 12 SP XY
X Y
n
4 12 48
(24)(40)
8 8 64 264
4
10 14 140
24
24 40 264
Earlier Subtopics 8.3.1 and 8.3.2 showed that Pearson r is computed by dividing
covariability of X and Y by variability of X and Y separately, where covariability
is the sum of products of deviation scores (SP). With these, the formula for
Pearson correlation can be written as:
Covariability of X and Y
r
Variability of X and Y separately
X M X M
X Y
SS X SS Y
SP
SS X SSY
Step 1: Find the deviation score of each variable by (a) computing the mean of
each variable, and then (b) subtract each variableÊs mean from each of
its scores.
E.g., Mx = (6 + 15 + 9 + 8 + 12 + 10) 6 = 10
6 10 = -4
E.g., (-4)2 = 16
E.g., SSx = 16 + 25 + 1 + 4 + 4 + 0 = 50
Table 8.4 shows the computing of the correlation coefficient for the milk
consumed and working memory study.
M = 10 M = 2.5
As you get a value from computing the Pearson correlation coefficient, you will
determine the strength of the correlation as stated earlier in Subtopic 8.2.3. The
closer the value to 1 (positive or negative), the stronger the relationship between
the two variables (Lund & Lund, 2013). Obtaining a value of μ1 indicates a
perfect relationship, that is, all the data scores falls on the line of best fit, and no
data deviates from the line. Examples of different relationships with different
correlation coefficients are shown in the Figure 8.7.
The study on the relationship between alcohol consumed and working memory
capacity shows a correlation coefficient of -0.82. As -0.82 is negative and near -1,
we can say that this study has a very large negative linear correlation.
SELF-CHECK 8.2
Person X Y
A 3 5
B 7 12
C 5 7
D 4 9
E 1 2
Person X Y
A 10 3
B 7 7
C 3 9
D 8 4
Thus, we should be aware that correlation does not explain why and how the
two variables are linked, but is only a pure connection. As stated by Gravetter
and Wallnau (2009),a correlation should not and cannot be interpreted as proof
of a cause-and-effect relationship between the two variables.
For example, suppose you are interested in finding the relationship between the
IQ and EQ. If you are using college students as your sample, your data may
represent only a limited range of IQ scores, as there are many students with
lower (e.g., below average) IQ scores who may not study up to college level. In
fact, the degrees of correlation may appear to be higher when you obtain a full
range of IQ scores in this study. An example is shown in the Figure 8.8.
Figure 8.8: Example of how full range of X and Y values gives a stronger correlation
compared to the restricted range of scores that gives a near zero correlation
Source:Gravetter&Wallnau, 2009
In brief, one should try to include a wide range of X and Y values (such as greater
numbers of a sample) in a study, be careful in interpreting the results, and not
generalise any correlation further than what is represented by the sample
(Gravetter& Wallnau, 2009).
For instance, as shown by Gravetter and Wallnau (2009), Figure 8.9 (a) shows a
sample of 5 with a correlation of -0.08, which indicates nearly zero correlation
between the two variables. However, when one extreme score is added to the
data set, it shows a correlation of 0.85, which indicates a strong, positive
correlation in the study. Without the outlier, we can say that there is no
relationship between the two variables, but, with the outlier, it turns out that
there is a strong and positive relationship between the two variables. Therefore,
Gravetter and Wallnau (2009) suggested us to look at the scatterplot instead of
simply looking at the numerical values obtained.
SELF-CHECK 8.3
2. Identify the outlier in the set of scores below and state the
potential problem(s) with regard to correlation.
X Y
14 45
10 36
16 50
11 42
4 72
Correlation r
Small About 0.10
Moderate About 0.30
Large About 0.50 or above
Aron et al.(2009), gave an example of their research with the idea that „doing
exciting activities together may cause people to be happier in their marriage‰.
Aron et al. (2009) reported that they have done many true experiments to support
the direction of causality for this idea, but the correlation remained low. This
issue tells them that there are many other factors that can lead to marital
satisfaction, which they did not include in the research or experiment (such as
doing exciting activities together). Hence, Aron et al. (2009) concluded that no
correlation can explain the whole story, and low reliability of psychology
measures may also play an important role in getting a small correlation.
On the other hand, there are psychologists who think that small correlations can
be important (Aron et al., 2009). For example, we seldom think that customer
satisfaction is affected by factors like environment. If it happens where there are
studies showing a minimal relationship between environment and customer
satisfaction, we are persuaded that the condition of the environment could play a
part in influencing customer satisfaction. In other words, small correlations may
not be relevant but may provide us related ideas or insight.
Last but not least, Cohen (1988) and Thompson (2007) proposed that the
significance of effects found in research should be compared with the effects
reported in previous similar research studies, instead of the CohenÊs conventions
(Aron et al., 2009).
Variable 1 2 3 4 5 6 7 8 9
8. Years of .492**
practice
9. Total
hours
playing
(Years
Hours)
ACTIVITY 8.3
Use your 3 main meals daily to compile your data. For each meal, note
its fats and its total calories. Identify if there is any relationship between
the two variables. State if it is a positive, negative or zero correlation.
Discuss the degree of the correlation and calculate the correlation
coefficient.
A scatterplot is a graph that shows the location of each pair of X-Y scores in
the data, which allows us to gain some idea of the direction of the
relationship between the two variables in a study.
There are a few patterns of correlation. Linear correlation occurs when the
dots in the scatterplot follow a straight line; curvilinear correlation occurs
when the dots follow a line pattern other than a simple straight line. When
dots do not follow any kind of line, it shows no correlation.
r
Covariability of X and Y
X M X M
X Y
The positive or negative sign of a correlation coefficient tells you the direction
of the linear correlation between the two variables.
Studies suggest that psychologists tend to see the correlation size as meaning
more of a strength between two variables compared to the actual scenario.
Nevertheless, small correlations may not be relevant but may provide us
related ideas or insights.
Aron, A., Aron, E. N., & Coups, E. J. (2009). Statistics for psychology (5th ed.).
Upper Saddle River, NJ: Pearson.
Gravetter, E. J., & Wallnau, L. B. (2009). Statistics for behavioral sciences (8th ed.).
Belmont, CA: Wadsworth, Cengage Learning.
Heiman, G. W. (2006). Basic statistics for the behavioral science (5th ed.). Boston,
MA: Houghton Mifflin Company.
INTRODUCTION
One of the psychologistÊs jobs is to make informed, and accurate predictions
about things, such as, how much a reading program is likely to help a particular
university student, or how well a teacher will perform in a new teaching job.
Statistical prediction also plays an important role in assisting research
psychologists to understand how outcomes of interest are influenced by various
factors (Aron et al., 2009). For instance, what are the lifestyle factors that predict
dementia in late adulthood? Learning statistical prediction in more detail will
help you to develop greater insight into other statistical topics and will prepare
you for the advance statistical techniques.
Y = bX + a
Y is the dependent variable (also known as criterion variable).
b is the slope. It determines how much the Y variable will change when X
increases. The direction can be either positive or negative.
X is the independent variable(also known as predictor variable).
a is the Y-intercept. a determines the value of Y when X = 0.
Let us take an example, using the number of children to predict the mothersÊ
stress levels. Suppose the regression constant a is 2, and regression coefficient b is
1. Hence, to predict the stress of a mother who has 4 children, you use the
formula:
Y = bX + a = 1 (4) + 2 = 6
As you can think, the number of children is not the only factor that predicts a
motherÊs stress level.
To see a step further, we are able to plot the graph with two points because the
two points completely determine a straight line. Nevertheless, it is advisable that
one computes and plots at least three points to get a definite graph (Gravetter &
Wallnau, 2009).
9.2.2 Regression
Regression is the statistical technique for finding the best-fitting straight line. The
straight line created from regression is known as regression line. This line is a
visual representation of how two variables are related (Aron et al., 2009).
SELF-CHECK 9.1
(a) 2.50;
(c) 4.89.
We want to minimise the error over the whole range of scores in our prediction.
In other words, we look for the smallest sum of errors. Sometimes the error
appears to be positive, where the rule is predicted too low and sometimes they
appear to be negative, where the rule is predicted too high. We wish to avoid the
errors that cancel each other out, that is, having the same amount of positive and
negative errors. We take each amount of error and square it and then we add up
the squared errors. Using squared errors have also various other statistical
advantages over alternative approaches. To evaluate if a prediction rule is good,
we calculate the sum of the squared errors we would make, using that rule (Aron
et al., 2009).
When choosing this linear prediction rule, we use what statisticians call the least
squares criterion. This means, the regression line gives the lowest difference in
sum of squared errors between the actual scores and the predicted scores on the
criterion variable.
The formulas for regression constant a and regression coefficient b are as follows:
a M Y (b )(M X )
b
X M (Y M
X y )
SS X
a Regression constant
b Regression coefficient
X M (Y M
X y ) Sum of products of deviation scores for all the
participants in the study
As you notice from the formulas, regression constant a can be obtained only
when we have the regression coefficient b value. Thus, we have to calculate the
value of b before calculating the value of a.
Steps to find the regression coefficient, b (see Table 9.1 for example):
Step 1: Similar to correlation (see Topic 8), find the deviation scores for each
variable by;
Step 3: Compute the sum of products of deviations by totaling all the products
of deviations scores.
Step 5: Sum up the squared deviation scores for variable X (predictor variable).
b
X M (Y M
X y )
SS X
Mx = 10 My = 2.5
X M
X (Y M y ) 28
b 0.56
SS 50
X
a = My ă (b)( Mx)
a = 2.5 ă (-.56)(10) = 8.1
Y = bX + a
Y= (-.56)(3) + 8.1 = 6.42
SELF-CHECK 9.2
X Y
4 6
6 8
7 3
3 7
SS X
(b )
SS Y
For example, referring back to the milk and memory example shown in Table 9.1,
we can compute the standardised regression coefficient by using this formula. As
it does not include the calculation of the value SSY, we may calculate by squaring
each deviation score (Y ă MY) and sum up the resulting values. Therefore, the
calculation can be seen as:
SS X 50 7.07
(b ) ( .56) ( .56) 0.82
SS Y 23.5 4.85
From this calculation, it shows us that for every standard deviation increase in
milk consumed, the predicted level of items remembered decreases by 0.82
standard deviations. With this standardised regression coefficient () value, you
could compare it with other studiesÊ values.
As you may notice from the example shown in Table 9.1, the value ( = -0.82) is
the same as the r value (r = -0.82) in Table 8.4. According to Aron et al. (2009), one
will get different and r values when scores computed from more than one
predictor variables are used to predict the scores on a criterion variable.
SELF-CHECK 9.3
ACTIVITY 9.1
From what you have learned about prediction so far, discuss with your
tutor in the forum on the issues that one may encounter in prediction,
including the standardised regression coefficient and hypothesis
testing.
Now, we would learn how to use more predictor variables. For instance, when
predicting occupational stress, we can examine the number of tasks assigned to a
person. Let us say we also know the number of tasks handled by the person
previously. Psychologist might be able to predict the personÊs occupational stress
more precisely by having this information. Multiple correlations can be made
when making the determination of the relationship between two or more
predictor variables and the criterion variable. Making these types of predictions
is called multiple regression (Aron et al., 2009).
Multiple regression linear prediction rule formula for five predictor variables is
(Aron et al., 2009):
In multiple regression, the relationship between the criterion variable and all the
predictor variables taken together is known as the multiple correlation coefficient
R. Because of the overlapping predictor variables, the multiple correlation is
usually smaller than the sum of the individual correlation coefficients. The
squared multiple correlation, R2, tells us how much the proportions of the
variance in independent variables are explained by the dependent variable(Aron
et al., 2009).
SELF-CHECK 9.4
1. In your own words, define multiple regression.
A regression coefficient tells us on how one predictor variable plays a part in the
prediction more than the other predictors (Goodwin, 2007). For example, if the
predictors are compared, the values for the standardised regression coefficients
ßs are 0.283, 0.467, -0.267. Meanwhile, for the rs values the predictors are 0.535,
0.108, 0.788. Then when reporting ßs, it can be said that the second predictor is
the most important. Next, for the rs, it shows that the third predictor is the most
important.
On the other hand, it is common to see the report of multiple regression results in
research articles. According to Aron et al. (2009), it is usually reported in table
form, together with the report of regression coefficients and overall
proportionate reduction in error. Table 9.2 showed an example of a regression
table written in APA 6th edition format.
A z score is the number of standard deviations the score is above the mean or
below the mean.
Regression is the statistical techniques for finding the best-fitting straight line,
which is known as the regression line.
Best prediction refers to the line that comes closest to the true scores on the
criterion variable.
Error occurs when the predictions ruleÊs predicted score is the actual score on
the prediction variable. We want to minimise the errors over the whole range
of scores in our prediction.
Aron, A., Aron, E. N., & Coups, E. J. (2009). Statistics for psychology (5th ed.).
Upper Saddle River, NJ: Pearson.
Gravetter, E. J., & Wallnau, L. B. (2009). Statistics for behavioral sciences (8th
ed.). Belmont, CA: Wadsworth, Cengage Learning.
INTRODUCTION
In this topic, you are going to learn some advanced statistical techniques, which
aim to give you some general idea of any statistical analysis in a research article.
Let us begin with an understanding of what you have learned so far. The main
part of the upcoming topic turns to a view of what lies ahead, in the future, as
you study more of the research on the basis of psychology. We will show you
how to make sense of the advance statistical procedures that are widely used in
psychology. Most research you will read as a learner of psychology uses one or
more methods that we have covered in this statistics course. It is also important
to know that you will be exposed to procedures that you have not learned, until
you take advanced statistic courses.
ACTIVITY 10.1
Before reading this topic, let us get some idea on carrying out a research.
Imagine that you are a researcher in the psychology field, and interested
to carry out a research. Identify the questions that you would ask
yourself in order to conduct a research and discuss them with your tutor
and coursemates. For example, you may discuss on your research idea or
title, steps to conduct a research, the details of your research (for
example, hypothesis, variables involved, measurement methods), and
other things.
Before running multiple regression analysis, we need to make sure that all
variables have a normal distribution. Distribution of variables that are not
normal (i.e., highly skewed or kurtosis variables or variables of substantial
outliers) can affect the relationship and the significance of the test (Sekaran,
2003). Correlation coefficient r shows strength between two variables, but it
cannot explain the differences in the dependent variables when the dependent
variable has been theorised to influence it at the same time (Sekaran, 2003). When
the variables jointly regress on the dependent variable, explaining the differences
among the individual correlations would fall on the multiple correlation. In other
words, multiple regression analysis will be the best method to assess the
significant relationships between independent variables and a dependent
variable.
In other words, Leary (2001) explained that stepwise regression analysis will stop
when:
(a) All the predictor variables are inserted in the equation, as each of them
make a unique contribution in predicting the outcome; or
(b) No predictor variable can uniquely predict any remaining variance in the
criterion variable.
For example, let us say a researcher is predicting work stress from relationship
with colleagues, job complexity, university or college CGPA, and job ambiguity,
with the following results:
* suppose job complexity is highly correlated with job ambiguity with r = 0.87
In the example, the researcher will enter job complexity in the equation as the
first step as job complexity has the highest correlation among the predictor
variables, which justifies the largest amount of variance in work stress. Next in
step 2, relationship with colleagues will be entered into the equation. Job
ambiguity is not included because it is highly correlated with job complexity. As
job complexity has justified the largest amount of variance in work stress that job
ambiguity can predict, job ambiguity then becomes a useless predictor. Thus, we
would include the next remaining variable that has the highest correlation, which
is the relationship with colleagues. After that, university/college CGPA will not
be entered into the equation because it does not enhance the prediction of work
stress. In the end of this example, only two predictor variables (i.e., job
complexity, and relationship with colleagues) are included in the regression
equation.
After that, the new predictor variables can be included into the equation as
follows:
(b) One-by-one; or
For example, when we know that a depressed person is less likely to socialise
with others. However, we cannot conclude that depression leads to social
withdrawal, as it can be confused with the presence of stigma. At this point,
hierarchical regression analysis can be used to partial out the variable that is not
contributing to depression. First, researchers may enter the presence of stigma as
10.3 RELIABILITY
Reliability refers to the measurement of consistency or regularity (Aron et al.,
2009). Put it another way, reliability of measure occurs when the respondents
answer the questionnaire in the same manner (i.e., same result) despite the
numbers administrated. On the other hand, if the answer varies, it shows that the
reliability of the question is not achieved. According to Nunally (1981), for this
reliability test, the Cronbach alpha value should be more than 0.7 to achieve
reliability of a question. Meanwhile, Sekaran (2006) stated that a reliable test
should have a Cronbach alpha value that is more than 0.6.
According to George and Mallery (2003), a commonly accepted rule of thumb for
describing internal consistency is as in the Table 10.1. A high value of reliability
means less error from the measurement tool, thus, making the measurement
more accurate and reliable. This is usually observed at the value rate of not less
than 0.60.
Source: CronbachÊs Alpha and Internal consistency (George & Mallery, 2003)
SELF-CHECK 10.1
To discuss in greater detail, the strength of the connection between each factor
and each measure differs, as one factor may have more influence on one measure
than the others (DeCoster, 1998). In factor analysis, the measures that have high
correlation (high factor loading) are likely to be influenced by the same factors
and vice versa (DeCoster, 1998).
Referring to the Table 10.1, Question 4 has the highest loadings on Factor 1. To
put it in other way, Factor 1 affects Question 4 the most as it has the highest
loadings, followed by Question 2 and 3. Similarly, Question 5 and 3 have the
most influence on Factor 2, while Factor 3 affects Question 5 the most.
According to Aron et al. (2009), factor loadings can be ranged from -1 to +1. The
loadings close to 1 denote a perfect (negative or positive) correlation with the
factor, loadings close to 0 suggest a weak influence of factor on the variable.
Therefore, a researcher would always try to get loadings close to 1. As stated by
Aron et al. (2009), psychologists often consider a variable to contribute
meaningfully to a factor only if it has a loading at least 0.30.
In path analysis, we use arrows or paths in path diagrams to identify and connect
the variables based on the theoretically based causal relationships that we would
like to test (Aron et al., 2009). According to Stage, Carter and Nora (2004), a
single-headed arrow points from cause to effect. A double-headed, curved arrow
indicates that variables are merely correlated; no causal relations are assumed.
Aron and his colleagues (2009) further stated that if the path diagram is a correct
description of the causal relationship among the variables, the path coefficient
tells you how much of a standard deviation change on the variable at the end of
the arrow is produced by a one standard deviation change in the predictor
variable at the start of the arrow.
Looking at the example in Figure 10.6, imagine that the research was aiming to
test a theory about what causes children to achieve in education. As shown in
Figure 10.6, we can see the cause-and-effect relationships pattern with path
coefficients among the variables. For instance, by looking at the arrows in the
path diagram, we get to know that parentÊs income is predicted by parentÊs age,
which in turn is not predicted to influence any other variable in the model (i.e.,
no arrows or paths from parentÊs income to any other variable). Nevertheless,
parentÊs education level is predicted to have influence on the childrenÊs
motivation in learning, which is one of the predicted causes of childrenÊs
intelligence, which in turn is predicted to affect childrenÊs educational
attainment.
As you probably noticed in the path diagram in Figure 10.5, all paths are
indicated withthe levels of significance by asterisks, as well, as path coefficients
that indicate the direction and degree of the predicted effects. Although path
diagrams show only the paths that are significant, it does not mean that the
researcher only tested these variables. In fact, the researcher tested all variables,
but include only paths that were significant in the diagram (Aron et al., 2009).
This is what we said earlier that a new model is developed via path analysis.
(d) Job satisfaction (M) is fully mediating or partially mediating the effect of
emotional intelligence (X) and perceived general health (Y).
Here is an example research with Figure 10.7 that was conducted by Smorti,
Cappelli, Guarnieri, Bergesio and Perfetto (2014), to show you how mediational
analysis is written or described in a research article.
The mediation regression analyses revealed that NYHA (New York Heart
Association) class positively predicted depressive symptom [ = 0.62; p < 0.001]
and negatively predicted avoidance [ = -0.68; p < .001]. Next, avoidance
negatively predicted depressive symptoms above and beyond NYHA class
[ = 0.52; p < 0.005]. When controlling for avoidance, the effect of NYHA class
on depressive symptoms was not significant [ = 0.25; n.s.]. The 34% of the total
variance in depression is explained by the indirect effect. A full mediation was
supported [z = 2.5, p < 0.01].
As the definition shows, there are a few advantages of having structural equation
modelling in a research (Aron et al., 2009):
As shown in the Figure 10.8, the measured variables are typically presented in
boxes, while the latent variables are shown in circles or ovals (Aron et al., 2009).
(a) The Arrows Go from the Latent Variables (in Circles) to the Measured
Variables (in Boxes)
The idea is that the latent variable is the underlying cause of the measured
variables, the measured variables being the best we can do to get the true
latent variable.
Last but not least, as stated by Aron and his colleagues (2009), the more
complicated the statistical procedures, the more likely we are to get misleading
results. For instance, group assignment (i.e., random assignment of group),
curvilinear relationship and restricted range can affect the results of the research.
SELF-CHECK 10.2
receives substitute). With this, we can see that the researcher manipulated the
variable by dividing it into an experimental group (i.e., group that receives
supplement) and a control group (group that receives substitute) in the study.
Variable like this is known as independent variable (IV).
On the other hand, dependent variable (DV) refers to the variable that the
researcher measures, which is expected to change according to the IV
manipulated by the researcher (Feldman, 2011). Simply put, DV is what the name
says, it „depends‰on the factor (i.e., independent variable) that the researcher
controls. Thus, the DV would be the effectiveness of the supplement (e.g., any
improvement in experimental group participantsÊ memory).
SELF-CHECK 10.3
(b) If you are the researcher, how would you set up the experimental
group and control group?
Student data were divided according to whether or not students had enrolled
in the final year PAL SSC, and t-tests (normal distribution assumed), which
were carried out to compare the Z-scores of each final examination component
between the two groups, for each year and all three year groups combined.
Analyses of covariance (ANCOVAs) were then performed on the combined
data using Year 4 results as a covariate to control confounding effects. ...
[Result] When final-year examination results for all years were amalgamated,
Z-scores for PAL tutors were approximately 0.2 higher for all components; this
difference was statistically significant. However, when Year 4 Z-scores were
controlled in ANCOVA, this difference was no longer significant. Subgroup
analyses for individual peer-tutor cohorts compared with non-PAL students
(n = 878) using ANCOVA showed that each peer-tutor cohort scored higher
than non-PAL students for each examination component, but the differences
were not significant. Exceptions included significant differences in SSO
Z-scores between students who peer-tutored on the March PAL SSC and non-
PAL students (Z = 0.425 and Z = 0.009, respectively; F(1,944) = 4.980,
p = 0.026), and in written examination, Z-scores between students who peer-
tutored on the December PAL SSC and non-PAL students (Z = 0.541 and
Z = 0.019, respectively; F(1,897) = 5.992, p = 0.015).
In short, the main difference between ANOVA and ANCOVA is the presence of
the letter „C‰ that refers to covariate which moderates the influence of IV on the
DV (Taylor, 2014). ANOVA includes only two variables which are IV and DV,
whereas ANCOVA involves an additional variable called covariate, which must
be a continuous variable (i.e., ratio data, interval data) (Statistics Solutions, 2016).
In terms of statistical procedures, we will also interpret the results in the same
basic way as others, where we will look at the testÊs significance level to decide
whether to reject or accept the null hypothesis (Aron et al., 2009). Hence, after
running MANOVA, it would tell us the differences between IV and DV groups
as a whole by showing us an overall F value, and significance level for the
differences among the IV groups. After that, as stated by Aron and his colleagues
(2009), if the groups differ significantly from DVs as a whole, we would need to
run univariate analyses of variance (i.e., individual analyses of variance) to see if
the IVs differ on any or all of the DVs.
Number of
Number of
Dependent
Statistical Association or Independent Any Variable
Variables (or
Technique Difference (or Predictor) Controlled?
Criterion
Variables
Variables)
Bivariate Association 1 1 No
correlation/
regression
One-way Difference 1 1 No
ANOVA, t-test
SELF-CHECK 10.4
1. Define covariate.
Nonetheless, as Aron and his colleagues (2009) stated, it would be best for us to
attend more or advance statistics courses to learn and keep ourselves updated
with the latest statistical techniques and procedures.
In conclusion, statistical methods keep changing and are invented from time to
time. It is common for psychologists to be puzzled with the numbers and
symbols they read in research articles. Therefore, be confident and open to learn.
With this attitude, you can master statistics (Aron et al., 2009).
ACTIVITY 10.2
Multiple regression analysis will be the best method to assess the significant
relationships between independent variables and dependent variables. It can
be done in stepwise method or hierarchical method. Stepwise multiple
regression analysis develops regression equation by including the predictor
variables one at a time, while hierarchical regression analysis include the
predictor variables based on their research hypothesis with reference to past
research or experiment.
Factor analysis helps to examine if the relationship patterns between all the
variables in a study can be explained by any unobserved variables (latent
variable or factors). A high factor loading means it is likely to be influenced
by the same factors.
Causal modelling sees beyond the correlation pattern between variables, and
checks further on the causal relations between variables whether they match
with certain theories. Path analysis illustrates these relationships with the
arrow directions, pointing from cause to effect. Mediational analysis is a type
of path analysis that examines whether a recognised causal relationship
between two variables is intervened by the mediator variable (M). Structural
equation modelling incorporates paths with underlying, unmeasured,
theoretical variables, allows a kind of significance test, and provides
measures of fit index.
Aron, A., Aron, E. N., & Coups, E. J. (2009). Statistics for psychology (5th ed.).
Upper Saddle River, NJ: Pearson.
Brodar, K. E., Crosskey, L. B., & Thompson, R. J. (2015). The relationship of self-
compassion with perfectionistic self-presentation, perceived forgiveness,
and perceived social support in an undergraduate Christian community.
Journal of Psychology and Theology, 43(4), 231-242.
Field, A. (2009). Discovering statistics using SPSS (3rd ed.). London: Sage
Publications.
Hyland, P., Shevlin, M., Adamson, G., & Boduszek, D. (2014). The organization of
irrational beliefs in posttraumatic stress symptomology: Testing the
predictions of REBT theory using structural equation modeling. Journal of
Clinical Psychology, 70(1), 48-59. doi: 10.1002/jclp.22009
Iwata, K., Furmedge, D. S., Sturrock, A. & Gill, D. (2014). Do peer-tutors perform
better in examinations? An analysis of medical school final examination
results. Medical Education, 48, 698-704. doi: 10.1111/medu.12475
Nunally, J.C. (1981). Psychometric theory (2nd ed.). New Delhi, Tata: McGraw-
Hill publishing Co Ltd.
Sekaran, U. (2003). Research method for business: A skill building approach (2nd
ed.). New York: John Wiley & Sons.
Sekaran, U. (2006). Research method for business: A skill building approach (4th
ed.). India: John Wiley and Sons.
Smorti, M., Cappelli, F., Guarnieri, S., Bergesio, F., & Perfetto, F. (2014).
Depression and cardiac symptoms among AL amyloidosis patients: The
mediating role of coping strategies. Psychology, Health & Medicine, 19(3),
263-272.
Stage, F. K., Carter, H. C., & Nora, A. (2004). Path analysis: An introduction and
analysis of a decade of research. The Journal of Educational Research, 98(1),
5-12. Retrieved from http://www.jstor.org/stable/27548053
Taylor, J. J. (August 11, 2014). Statistical soup: ANOVA, ANCOVA, MANOVA, &
MANCOVA. Retrieved from http://www.statsmakemecry.com/smmcthe
blog/stats-soup-anova-ancova-manova-mancova
Yuan, L., Tan, X., Huang, C., & Zou, F. (2014). Mediating effect of job satisfaction
on the relationship between emotional intelligence and perceived general
health. Social Behavior & Personality: An International Journal, 42(7), 1057-
1067.
OR
Thank you.