Download as pdf or txt
Download as pdf or txt
You are on page 1of 66

Psychology 253

Data Analysis in Psychology


Why?
~ Understand psychology phenomena
~ Prevalence
~ Relationships between variables
~ Test efficacy of interventions

Chapter 1: Introduction to Statistics


Scales of Measurement
~ Any data collection requires that we make measurements of our observations.
~ It involves...
o The categories used to measure a variable, makes up a scale of measurement.
o The relationships between these categories determines the different types of scales.
~ Measurement assigns individuals or events to categories
o The categories can be names, such as male/female or employed/unemployed
o They can be numerical values, such as 68 inches or 175 pounds
~ The complete set of categories makes up a scale of measurement
o Relationships between the categories determine different types of scales
There are 4 different types of scales…
Scale Characteristics Examples
Nominal •Label and categorize •Gender
•No quantitative distinctions •Diagnosis
•Experimental or Control
Ordinal •Categorizes observations •Rank in class
•Categories organized by size or •Clothing sizes (S,M,L,XL)
magnitude •Olympic medals
Interval •Ordered categories •Temperature
•Interval between categories •IQ
of equal size •Golf scores (above/below
•Arbitrary or absent zero point par)

Ratio •Ordered categories •Number of correct answers


•Equal interval between •Time to complete task
categories •Gain in height since last
•Absolute zero point year

~ This material is arguably in the “Top Ten Most Important” concepts the students will encounter in the
study of statistics and may merit identifying it as such.
~ The word nominal means having to do with names or labels.
o It involves classifying individuals or events into categories.
o They all have different names but are not related in any way.
o For example: if you were measuring academic majors for a group of university students, those
majors would be classified according to categories such as: psychology, sociology, business
science, biology etc. so each student in that group would be classified in a category.
~ An ordinal scale consists of categories which are organised in a sequence.
o So, measurements are ranked in class.
o Sizes etc.
o For example: in a psychology module, one could rank students according to who came first,
second, third or fourth in the course in terms of academic achievement. The fact that
categories form an ordered sequence means that there is a relationship between categories.
~ An interval scale consists of ordered categories that are all intervals of exactly the same size.
o So, it is characterized by equal intervals between scale units.
o Basically, the difference between two values are meaningful.
o For example: Person A scored 65% in her psychology 253 exam and person B scored 80% in
her psychology 253 exam. We can also then say that person B scored 15% higher than person
A and we can go further and say that it is equal to Person C scoring 70% and Person D scoring
85% (the difference between both sets of scores is equal). When thinking of this in terms of a
scale, the zero point of the scale is arbitrary, in other words it has no zero point.
~ The ratio scale is basically an interval scale with and added characteristic.
o It has an absolute zero which means that a score of zero indicates none of the variable being
measured.
o For example: number of correct answers on a student’s psychology 253 exam. It can either be
any number of correct answers as well as zero correct answers.
Three Data Structures, Research Methods, and Statistics
Data Structure I:
o Descriptive research (individual variables)
o One (or more) variables measured per individual
o “Statistics” describe the observed variable
o May use category and/or numerical variables
~ Data structure 1: One or More Separate Variables Measured for Each Individual:
~ Descriptive Research:
~ This involves measuring one or more separate variables for each person or participant.
~ The intention here is to simply describe as the title says.
~ Think of this example
Individual Number of hours Number of hours Number of hour
exercise in a day sleeping in a day studying in a day

A 2 6 4

B 1 7 3

C 3 4 5

D 4 8 2
~ So in this table we can see that it speaks to how many hours an individual exercises, sleeps and
studies in each day.
~ Here it is different for each person, and we are simply describing the variables by saying that
Person A exercises for 2 hours a day, sleeps for 6 hours and studies for 4 hours each day, and so
on with the rest of the individuals described here.
Relationships between variables
~ Is very important.
o Two (or more) variables observed and measured
o One of two possible data structures used to determine what type of relationship exists
~ Most research aims to examine whether there is a relationship between variables. For example:
~ Is there a relationship between number of hours spent studying a day and the results on a test?
~ Is there a relationship between number of hours spent studying for psychology 253 exam and the
results of the exam?
~ To establish whether there is a relationship – we first have to make observations of the variables.
Data Structure II:
o The correlational method
o One group of participants
o Measurement of two variables for each participant
o Goal is to describe type and magnitude of the relationship
o Patterns in the data reveal relationships
o Non-experimental method of study
~ Data Structure 2 involves the correlational method.
~ So, this is observing one group with two variables being measured.
~ So we just spoke about examining the relationship between variables and this can be done using two
variables.
~ Simply put, we can measure two variables for each individual.
~ You have examples in your textbook but I want to focus on another example here:
~ In the correlational method two different variables are observed to determine whether there is a
relationship between them.
~ Example
STUDENT NUMBER OF HOURS ACADEMIC
SPENT ON SOCIAL PERFORMANCE
MEDIA (Results on a test)
A 5 50%

B 6 40%

C 4 60%

D 2 70%

o This table shows information on 4 students in relation to the amount of time they spend on
social media and the results they score on a test.
o So, you can see that person C spends 4 hours on social media and scored 60% on her test
o Whereas person B spends 6 hours on social media and scored 40% on his test…
o draw scatterplot with whiteboard function.
Figure 1.5 Data structures for studies evaluating the relationship between variables
o One of two data structures for studies evaluating the relationship between variables.
o Note that there are two separate measurements for each individual (Facebook time and academic
performance).
o The same scores are shown in table (a) and graph (b).

Correlational Method Limitations


o Can demonstrate the existence of a relationship
o Does not provide an explanation for the relationship
o Most importantly, does not demonstrate a cause-and-effect relationship between the two
variables
o So, as with any technique there may be some limitations.
o Regarding the correlational method that we just heard, we can deduce that the results of a
correlational study can demonstrate the existence of a relationship between two variables, but
they do not explain the relationship.
o More noteworthy is that a correlational study cannot demonstrate and cause and effect
relationship between the two variables.
o So, as I mentioned in the example before the table shows that the more hours spent on social
media resulted in a lower percentage scored on a test, however there could also be other factors
involved that influenced that student or individual to score low on his or her test that day.
o Hence, we can’t conclude to say that if you spend less time on social media your academic
performance would increase…
Data Structure III:
o Comparing two (or more) groups of scores
o One variable defines the groups
o Scores are measured on second variable
o Both experimental and non-experimental studies use this structure
~ Instructors may wish to introduce the term “quasi-experimental” in this section.
~ The third data structure is: Comparing two or more groups of scores: Experimental and Non-
experimental methods:
~ In this data structure, you use one variable to define groups and then measuring the second variable
to obtain scores for each group.
o Example: You have two groups… As mentioned in the text book: 1 group is exposed to a
violent video game and another group exposed to a non-violent video game. The results can
be seen in the table on the next slide…
Figure 1.6: Data structure for studies comparing groups]

o The second data structure for studies evaluating the relationship between variables.
o Note that one variable is used to define the groups and the second variable is measured to
obtain scores within each group.
o So here you are assessing two variables, type of video game (violent and non-violent) and the
second variable - aggressive behavior is measured to obtain the scores as seen in this table
above.
Experimental Method
~ The experimental method is systematic and scientific approach in which the research manipulates one
or more variables.
~ So, one variable is manipulated and the other is controlled.
o For example: administering a diet supplement to one group and the other group receives a
placebo; there weight-loss is being measured.
~ Goal of experimental method
o To demonstrate a cause-and-effect relationship
o This requires manipulation by the researcher changing its value from one level to another.
~ Manipulation
o The level of one variable is determined by the experimenter
o So relating back to the previous example about violent video games and non-violent video
games.
o The research then manipulates by giving one group of boys a violent video game to play and
the other group non-violent video game.
o The second variable as mentioned is the variable that is measured and in this example it is
whether the manipulation had an effect or not.
~ Control rules out influence of other variables
o Participant variables
o Environmental variables
o Regarding control, this is where the researcher makes sure that no external factors influence
the relationship.
o So in the example of the diet supplement where one group receives the diet pill and the other
group receives the ‘fake’ diet pill (placebo), the researcher then makes sure that no extraneous
factors influence the relationship by ensuring maybe the both groups do the same exercise
routine and eat the same foods etc.
Experimental Method: Control
~ Within the experimental method, there can be controlled and experimental conditions.
~ Individuals in a controlled condition do not receive the experimental treatment (for example, the
group of boys who played the on-violent video games or the one group of people who received the
placebo diet pill.
~ Methods of control
o Random assignment of subjects
o Matching of subjects
o Holding level of some potentially influential variables constant
~ Control condition
o Individuals do not receive the experimental treatment
o They either receive no treatment or they receive a neutral, placebo treatment
o Purpose: to provide a baseline for comparison with the experimental condition
o The purpose of such a condition is to provide a baseline for comparison.
~ Experimental condition
o Individuals do receive the experimental treatment
o Individuals in the experimental conditions do receive the experimental treatment (i.e the
violent video games or the diet pill)
Independent/Dependent Variables
~ Independent variable is the variable manipulated by the researcher
o Independent because no other variable in the study influences its value
~ Dependent variable is the one observed to assess the effect of treatment
o Dependent because its value is thought to depend on the value of the independent variable
~ So in our example the independent variable would be the amount of violence (so the violent vs non-
violent video games) and the dependent variable is the level of aggressive behaviour.
Non-Experimental Methods
~ Non-equivalent groups
o Researcher compares groups
o Researcher cannot control who goes into which group
~ Pre-test / Post-test
o Individuals measured at two points in time
o Researcher cannot control influence of the passage of time
~ Independent variable is quasi-independent
~ There are also non-experimental methods.
~ The first being no-equivalent groups:
o Any study that follows a scientific requirement such as the previous examples are known as
experimental studies.
o And as such there are also designs that are no-experimental but still examine the relationships
between variables.
~ A pre-test and post-test study uses time such a s before and after to create groups of scores. Lets
see the examples…
Two examples of non-experimental studies: Figure 1.7
~ Two examples of non-experimental studies that involve comparing two groups of scores.
~ In (a), a participant variable (gender) is used to create groups, and then the dependent variable
(verbal score) is measured in each group.
~ In (b), time is the variable used to define the two groups, and the dependent variable (depression)
is measured at each of the two times.
Chapter 2: Frequency Distributions
Frequency Distributions
~ A frequency distribution is
o An organized tabulation
o Showing the number of individuals located in each category on the scale of measurement
~ Terminology associated with frequency distributions is one of the least “standardized” across
disciplines and texts students might encounter.
~ Instructors may wish to emphasize the importance of being precise with the terms provided by the
text authors, but also be aware that terms may differ in other texts or courses.
~ Can be either a table or a graph
~ Always shows:
o The categories that make up the scale
o The frequency, or number of individuals, in each category
~ I am not going to go into any definitions as you as the student needs to make sure you go through
every concept and make sure you become familiar with the terminology used in statistics.
~ This is crucial…
o So we know that a frequency distribution is basically data organized in a table that can explain
something…
o It can also be distributed onto a graph like we saw in the previous podcast on data structures.
o But here you will learn the statistical input of data in tables or graphs. Like I said before
practice makes perfect, so go through your textbook and practice using the Learning Check
questions.
o So you can know more or less how you may be examined or what a certain questions asks of
you.
o In any frequency distribution it indicates the categories which make up the scale as well as the
frequency or number of individuals in each category.
Example
X f
8 9 8 7 10 9 6 4 9 8
7 8 10 9 8 6 9 7 8 8 10 2

9 5

N= (number of scores 8 7

20) 7 3
X= score
6 2
f= number of times
5 0
scores occurred.
4 1

~ Here you have a set of scores right… there are 20 scores.


~ The sum of the scores are denoted with the N symbol.
~ We say here N equals 20 as there are 20 scores.
o In statistics, a frequency distribution takes a set of scores (and usually we organise the scores
from highest to lowest) but it takes that set of scores and places it in a table.

ΣX = N
X f

5+4+4+3+3+3+2+ 5 1

2+2+1= 29 4 2

3 3

ΣX = 29 2 3

1 1

ΣX2= 97
o see this table here… we have 20 scores but we can see that some scores present more than
once so we can group it as follows…
o we can see that N= sum, X=score and f=frequency (number of times score occurred). Again
here you must basically memorize these symbols because they are used throughout stats.
§ So these set of scores are organised in the table under the column X (excluding the
scores that occur more than once) and remember it must always be tabulated in order
from highest to lowest.
§ Once you organise your scores from highest to lowest you look at each number and
see how many times a score of 10 occurred, a score of 9 occurred and so on…
§ Here you actually count how many timesa score of 10 occurred etc..
§ You can now see clearly the scores in the table. 2 people scored the highest score
which is 10 and you can also see that no one scored 5 but it is included in the table.
§ Remember the 4 different scales that we spoke about previously, with an ordinal,
interval and ratio scale the categories are listed in order rom highest to lowest.
Frequency Distribution Tables
~ Structure of frequency distribution table
o Categories in a column (often ordered from highest to lowest but could be reversed)
o Frequency count next to category
~ Σf = N
~ To compute ΣX from a table
o Convert table back to original scores or
o Compute ΣfX
~ As mentioned, the categories are ordered from highest to lowest and they are displayed under the
column X and the number of times that score appeared is found in the column f.
~ But now we want to know the total numbers of scores in the distribution and this is denoted by the
symbol above.
o to calculate the sum of scores in the distribution you need to look at both columns.
o Let’s look at the next example…
Calculating the scores in a Frequency Distribution
Consider this table…
X f ΣX

5 1 5

4 2 8

3 3 9

2 3 6

1 1 1

ΣX = 29

~ You see a set of scores and how many times the score appeared which is the frequency.
~ So to calculate the sum of the scores we do the following:
o Remember I said in the previous slide you need to consider both columns when calculating
the sum…
~ You will write out each score including the amount of times each score appeared:
~ 5+4+4+3+3+3+2+2+2+1= 29
~ To get the sum on the frequency squared you square each score and add the squared values = 97
This you do on a calculator…
~ Another way to get the sum of X and input in in a table is as follows…
~ Let’s input into the table together…
Proportions and Percentages
Proportions
• Measures the fraction of the total group that is associated with each score
f
proportion = p =
N
• Called relative frequencies because they describe the frequency ( f ) in relation to the total
number (N)
Percentages
• Expresses relative frequency out of 100
f
percentage = p(100) = (100)
N
• Can be included as a separate column in a frequency distribution table
~ The ability to quickly and comfortably convert between fractions (proportions), decimal fractions
(relative frequency), and percentages is fundamental to success in this course.
~ Some students struggle with reconciling the fact that although these are three distinct metrics, they
all point to the same “deep” meaning.
~ There are other measures that describe the distribution of scores which we can also interpret.
~ These are Proportions and Percentages.
§ Proportion measures the fraction of the total group associated with each score so for
example we see in the previous table that two people scored 4 so we can say that 2
out of 10 people had 4, this is then demonstrated in this form above.
~ Researchers also use percentages to describe a distribution, and this can be done by first finding the
proportion and then multiplying that by 100.
§ Use whiteboard to show proportion and percentage...
Example 2.4: Frequency, Proportion and Percent

Grouped Frequency Distribution Tables


~ If the number of categories is very large, they are combined (grouped) to make the table easier to
understand
~ However, information is lost when categories are grouped
o Individual scores cannot be retrieved
o The wider the grouping interval, the more information is lost
~ Today’s podcast will cover grouped frequency tables…
~ In the previous podcast we had a distribution of 20 scores etc, but what about the distributions with
very large categories? This is when grouped frequencies come in…
o when we have a very large range of values, we cannot list all scores in a table we then turn to
grouped frequencies.
o Our aim is to always gain a simple organised picture of a distribution, that is why we use
tables.
o so a grouped frequency can be done by grouping scores into intervals and then listing those
intervals in a table instead of each individual score.
§ So for examples we can group the number of scores of students who had marks in the
80s and 90s etc… this table is called a group frequency distribution table.
§ We can see an example in the next slide.
Example
~ In this table you can see that instead of listing individual scores.
~ When you have large distributions you can group them as follows.
~ Deciding on how to group these scores involves following some guidelines.
~ So you will have the full set of score in front of you and then you create the table with the intervals
based on the scores range.
~ Then you will look at the scores and see how many scores from the list fit in to each interval.
~ So for example, 3 of the scores ranged between 90 to 94. (The actual scores were 91; 94; 93).
X f
90-94 3
85-89 4
80-84 5
75-79 4
70-74 3
65-69 1
60-64 3
55-59 1
50-54 1

“Rules” for Constructing Grouped Frequency Distributions


~ Requirements (Mandatory Guidelines)
o All intervals must be the same width
o Make the bottom (low) score in each interval a multiple of the interval width
~ “Rules of Thumb” (Suggested Guidelines)
o Ten or fewer class intervals is typical (but use good judgment for the specific situation)
o Choose a “simple” number for interval width (e.g., 2,5,10)
~ This slide shows a recommended treatment for the four guidelines presented in the text.
~ The text presents it slightly differently by indicating these are “guidelines” rather than absolute
requirements.
~ However, violating guideline 4 distorts the information conveyed and violating guideline 3 makes it
much more difficult to assimilate the information conveyed by the table.
~ Consequently, each instructor should clarify expectations for her class: are these “guidelines” or
“rules?”
~ So here we want to give you some guidelines in terms of grouping frequencies.
~ Firstly, the grouped frequency distribution table should have about 10 intervals (Go back to table and
show them).
~ Remember the whole purpose of a frequency distribution table to is help the researcher or reader to
see the data in an organized manner.
~ Second guideline is that the width or size of the intervals should have a simple number such as 4 or 5
or 10 or lower even like 2.
~ Here you need to think what would be easy to count in so 5s or 10s seem relatively easy to count.
~ Thirdly, the bottom score in each class interval should be a multiple of the width.
o For example, if you are using a width of 10 points then the intervals should start with 10, 20,
30 40, etc.…
~ The fourth guideline is that all intervals should be the same width.
~ This is to ensure that any score will fit or belong into these intervals.
o refer to the table example: here you can see that the width of 5 was sufficient to group these
scores.
o remember the wider the intervals are the more information is lost so you want to keep it as
concise ad possible because in a group frequency table you can’t tell exactly what the score is.
The shape of a Frequency Distribution
~ We get an idea of the shape of the distribution
~ Can enhance the understanding of the nature of the data
~ There are many different ways you can input data from tables onto graphs, you would have come
across the various types of graphs such as bar graphs, scatterplots etc.
~ But what is almost more important is what we can see and explain from looking at a graph.
~ Here, the shape of the frequency distribution on the graph is very important.
~ There are three characteristics that completely describe any distribution and these are:
1. Shape: we classify the graphs according to the shape, so whether they are symmetrical or
skewed.
2. Central Tendency: this measure where the centre of the distribution is located
3. Variability: this is relating to the degree of whether the scores are spread over a wide range or if
they are clustered at one point.
Figure 2.9 – IQ Population Distribution Shown as a Normal Curve

~ The population distribution of IQ scores: an example of a normal distribution.


~ In this graph you can see something that is called a normal curve.
~ A normal curve is where the greatest frequency is in the middle and smaller frequencies move to
what is called the extreme ends.
~ In this example of IQ scores distributed on this graph, it then tells us that majority of people’s IQ
scores that were tested are found in the middle.
~ The extreme scores then lie on either ends.
~ It is extreme for the fact that it shows few people had a very low IQ score of less than 70 and on the
other extreme we can see that few people had IQ scores of higher than 130.
Figure 2.10 Distribution Shapes
~ Examples of different shapes for distributions.
~ As I mentioned in the first slide, we identify shape according to symmetry and skewness.
~ In a symmetrical distribution it is possible to draw a straight line through the middle of the graph so
that one side is identical to the other… (show them the line…)
~ Then, with skewness, this is also important in that it tells us where scores tend to pile up at one point
of the graph and taper off gradually towards the other end.
~ This is what we call either positively or negatively skewed.
~ Any time you see a skewed distribution to the right side of the graph then it is a positively skewed
distribution because the tail of the graph points to the positive on the x-axis.
~ Any time you see a skewed distribution the left-hand side of the graph then it is termed a negatively
skewed distribution.
~ For example, if you have a group of students who wrote a difficult exam, you will know that majority
of the scores would be low with only a few individuals scoring high marks this would produce a
positively skewed distribution.
~ In a similar instance where students write and easy exam, majority of the students will score high and
few score low and this in turn shows a negatively skewed distribution.
Example: Negatively Skewed

Example: Positively Skewed


Chapter 3: Central Tendency
Figure 3.1: Locate Each Distribution “Center”
~ Three distributions demonstrating the difficulty of defining central tendency. In each case, try to
locate the “center” of the distribution.
~ This slide presents a good opportunity for a discussion of how central tendency should be
characterized and how the same measure of central tendency may not be suitable for different
distributions
~ We know that central tendency is a statistical measure to determine which single score defines the
center of any distribution which then is the most representative of the entire group of scores.
~ This is easily seen on a normal distribution, but it becomes tricky when we have positively and
negatively skewed distributions.
~ In the las podcast we spoke about the different shapes of distributions but now we are also able to
calculate the mean of distributions.
~ You can see on the graphs here that there are different distributions.
~ A normal, a negatively skewed distribution and the bottom one which is normal but has two sets of
distributions.
~ To accurately measure central tendency we can calculate the mean, median and mode.

Central Tendency Measures


~ Figure 3.1 shows that no single concept of central tendency is always the “best”
~ Different distribution shapes require different conceptualizations of “center”
~ Choose the one which best represents the scores in a specific situation
~ As mentioned now, the goal of central tendency is to provide the best representation of scores in a
distribution. Now we can look at calculating the mean…
The Mean
~ The mean is the sum of all the scores divided by the number of scores in the data
~ Population:

µ=å
X
N
~ Sample:

M=
åX
n
~ Instructors my wish to have students compare and contrast these two formulas.
~ The mean is basically the average…
~ To calculate the mean we all all the scores and divide by the number of scores in the data.
~ The formula to calculate the mean is shown here.
~ There are two different ways to calculate the mean firstly in terms of the population and secondly for
a sample.
~ Here you can spot the slight differences between the two formulas.
~ For the formula to calculate the mean for the population it is as follows:
~ So this symbol represents the population mean and it is calculated by adding all scores and dividing
by the number of scores. So the sum of X over N.
~ For the sample mean the symbol is M=the sum of x over the number of scores…
~ So if you see a mean being represented by and M then you are dealing with a sample mean.
~ If it it being represented by the U symbol, then you are dealing with a population mean.
~ Lets looks at an example:
Example:
~ Calculating the population mean:
5, 6, 7, 3, 4
~ N= 5 scores
~ Add up all scores first: 5+6+7+3+4= 25
~ The mean is: 25
o So remember we first count how many scores there are, we see that there are 5 scores so N=5
o Then we add up all the scores which gives us 25
o Now to calculate the population mean we say pop mean = 25 divided by 5 =5 therefore our
population mean is 5.
~ Calculating the sample mean:
3, 6, 5, 3, 4
~ N= 5 scores
~ Add up all scores first: 3+6+5+3+4= 21
~ The mean is:
~ Gym attendance per week
The Weighted Mean
~ Combine two sets of scores
~ Three steps:
o Determine the combined sum of all the scores
o Determine the combined number of scores
o Divide the sum of scores by the total number
of scores
Overall Mean =

M=
åX +åX 1 2

n1 + n2
~ So what if we want to now know the mean of two sets of scores?
~ This is where we use the weighted mean formula.
~ So this formula is the Overall Mean equals the sum of scores for group 1 plus the sum of scores for
group 2 divided by the number of scores for group one and the number of scores for group two.
~ Now let’s see this in an example:
Example:
~ Hours of studying for a test per day:
~ Group 1: 3, 4, 5, 2, 6
~ Group 2: 2, 5, 3, 1, 5
~ So here we have a set of scores which indicate the number of hours a day two groups of 5 people
each, study for a test.
~ So we want to find out what the average or mean hours both groups studied for a test.
~ Group 1 hours are 3,4,5,2,6
~ Group 2 hours are 2,5,3,1,5
~ So let’s follow the steps we just spoke about for calculating the weighted mean:
~ Firstly lets add the total scores for each group: Group 1: 20 so sum of x =20
~ Group 2: 16 so the sum of x=16
~ And we know that there are a total of 5 scores per group so now lets write this out…
Computing the Mean from a Frequency Distribution Table
Quiz Score (X) f fX
10 1 10
9 2 18
8 4 32
7 0 0
6 1 6
Total n = Σf = 8 ΣfX = 66

M = ΣX / n = 66/8 = 8.25

~ We can also compute the mean from looking at a table.


~ So if a set of scores are inputted in a table, the calculation is usually easier to look at the column for X
and the column for F.
~ So we can see that there is one 10, two 9s, four 8s and one 6.
~ Our step is to add them all up so 10+9+9+8+8+8+8+6= 66
~ Next we can see that we have a total of 8 scores…
~ So our mean is inputted as follows: M= Sum of X divided by N which is 66 divide by 8= 8.25 therefore
the weighted mean is 8.25
The Median
~ The median is the midpoint of the scores in a distribution when they are listed in order from smallest
to largest
~ The median divides the scores into two groups of equal size
~ The second measure of central tendency is the median.
~ This has to do with finding the midpoint in a distribution.
~ So if you receive a list of scores it first needs to be organised from smallest to largest.
~ You look for the score that lies directly in the middle of the distribution of scores.
Example 3.7: Locating the Median (odd n)
~ Put scores in order
~ Identify the “middle” score to find median
3 5 8 10 11
“Middle” score is 8 so median = 8
~ So here for example you have a set of scores, you then order the scores from lowest to highest.
~ Now we look for the score that lies in the middle and has equal number of scores on both ends.
~ You can see that there are two scores on either side of this set of scores which then allows us to see
that the middle score is 8.
~ But what if we have a list of scores that is an even number?
Example 3.8: Locating the Median (even n)
~ Put scores in order
~ Average middle pair to find median
1 4 5 7 9
(4 + 5) / 2 = 4.5
~ Instructors may also want to point out the 4.5 is exactly ½ the distance between 4 and 5 on interval
or ratio measurement scales.
~ So you follow the same process in listing the scores from lowest to highest.
~ You can see we have 6 scores in total which is an even number.
~ We would then look at the two scores that lie in the middle.
~ Now we take those two score and add them up and then divide it by 2. this will give us the median.
So 4+5=9 and 9 divide by 2= 4.5
~ Therefore 4.5 is the median for this list of scores.
Finding the Precise Median for a Continuous Variable
~ A continuous variable can be infinitely divided
~ The precise median is located in the interval defined by the real limits of the value.
~ We must determine the fraction of the interval needed to divide the distribution exactly in half.
number needed to reach 50%
fraction =
number in the interval
~ Students have sufficient difficulty with this concept to justify developing another example similar to
the one the authors provided in Figure 3.4.
~ So what if we also want to find the precise median for a continuous variable?
~ This can be a little tricky.
~ We know a continuous variable consists of categories that can be split into fractional parts.
~ We must determine the fraction of the interval needed to divide the distribution in exactly half.
~ Such as time for example can be split into seconds, 10th of a second and hundredth of a second.
~ So by splitting it into fractional parts we can find the median by locating the precise point that
separates the bottom 50% of the distribution by the top.
~ So to calculate the fraction we use this formula:
~ Fraction= number needed to reach 50% divided by the number in the interval
~ Lets see an example…
Example
2, 3, 4, 4, 4, 5
~ Between first and second 4. three of the numbers are below and three are above right as that is the
definition if the median, so I want to show you that there is a number between 4 and 4.
~ So lets use an example here.
~ We ask 6 people how far it is that they have to walk from their house to the nearest shop in their
neighborhood…
~ So the first person said 2km, 3km, 4km etc.
~ Sometimes when you ask people how far they live they don’t always give you the precise number or
distance.
~ So lets put these numbers on a number line so that we can see what we are talking about
The Mode
~ The mode is the score or category that has the greatest frequency of any score in the frequency
distribution
o Can be used with any scale of measurement
o Corresponds to an actual score in the data
~ It is possible to have more than one mode
~ Some instructors may wish to extend the discussion of bimodal (and multimodal) distributions to
include the concept of major and minor modes to recognize that even when there is technically only
one “most frequently occurring” score, it might be helpful and useful to report more than one mode
to help better characterize the distribution.
~ The mode is basically the score that appears the most or has the greatest frequency…
~ The mode is used because it can determine the typical or most frequent value for any scale of
measurement, so for example we can say a black pen is the mode for students in PSY253 class as
most of the students write with a black pen.
~ This is determined from the amount of test or exam papers we receive.
Example
1,3,4,2,3,3,5,6,5,6,6,7,8,6,7,7,9,10,10,9,10
Mode: 6 as it appears four times
Shop f

Checkers 32
N=100 scores
Mode= 32
Woolworths 18

Pick ‘n Pay 28

Shoprite 22

~ So a basic example is this set of scores, you can see that 6 appears the most (4 times) hence 6 is the
mode.
~ Lets take another example by looking at the table above, you can see the different stores.
~ So for example a sample of 100 students on campus were asked which store they shop at for their
groceries, the most frequent answer is Checkers as 32 out of 100 students said they shopped at
Checkers.
~ It is also important to note that sometimes there can be two modes in a frequency.
~ Bimodial: is a distribution with 2 modes and a distribution with more than 2 modes is called
multimodial.
Chapter 4: Variability
Defining Variance and Standard Deviation
~ Most common and most important measure
of variability is the standard deviation
o A measure of the standard, or average, distance from the mean
o Describes whether the scores are clustered closely around the mean or are widely scattered
~ Calculation differs for population and samples
~ Variance is a necessary companion concept to standard deviation but not the same concept
o The twin concepts of variance and standard deviation are among the most challenging
concepts in a basic statistics course to communicate and to learn.
o Instructors will almost certainly want to invest special care in the preparation of materials to
help communicate these very difficult concepts.
o So in today’s podcast we wil focus on defining variance and standard deviation.
o The standard deviation is the standard or average distance ffrom the mean, its going to give us
a description of whether scores are close to or cluster together on the mean or widely
scattered out.
o Now we want to note that here too lik when calculating the mean there were slightly different
formulae, so here to calculate the SD of the population and samples are different too.
o Please ensure to familiarize yourself with the formulae in your text book as it also clearly
differentiates between sample and population formulae.
o Now Variance is a somewhat companion concept of SD. BUT it is not the same concept.
o So lets go through these formulae.
~ Step One: Determine the deviation
~ Deviation is distance from the mean
Deviation score = X − μ
~ Step Two: Find a “sum of deviations” to use as a basis of finding an “average deviation”
o Two problems
§ Deviations sum to 0 (because M is balance point)
§ If sum always 0, “Mean Deviation” will always be 0.
o Need a new strategy!
~ Having students try to come up with an intuitive method for developing a measure of variability
based on deviation scores is a great way to get them thinking about what a dead end strategy
averaging deviations is (because, of course, the average of deviations from the mean must be 0).
~ Several teams working on it in a classroom exercise often results in a valuable insight about the issue
(averaging absolute value of deviations) and might produce the one we use—squaring the deviations
to eliminate the negative values.
~ When we calculate the SD we basically asking how different the scores are from one another. So
deviation as mention means distance.
~ So for example: think of a psy 253 exam where the average grade is 70% but your score is 80%, so
essentially your score deviates from th mean by positive 10 points and likewise if you score below 80
you areminus 10 points from the mean.
~ So we can calculate for everyone in the class and calculate the deviation score for each person and at
the end we can calculate the average score for all deviations in the class.
~ This is essentially what SD is. It tells us the average distance of the sample from the pop mean.
Example
Finding the deviation score:
Deviation =x-µ µ=50 x=53
=53-50
Deviation score=3
Deviation =x-µ µ=50 x=45
=45-50
=-5
To get rid of the (-) you have to Ö (square) each and every score and then calculate the variance.
~ Step Two Revised: Remove negative deviations
o First square each deviation score
o Then sum the squared deviations (SS)
~ Step Three: Average the squared deviations
o Mean squared deviation is known as “variance”
o Variability is now measured in squared units
Population variance equals mean (average)
squared deviation (distance) of the scores
from the population mean
o The concept of sum of squared deviations (SS) is absolutely vital to efficient understanding of
the statistical tests presented in the remainder of the text.
o The authors have reduced the computational complexity and the cognitive load required of
students—contingent upon grasping and retaining the concept of SS presented in this chapter.
o The authors also lay the foundation for efficiently learning the fundamentals of ANOVA—
contingent upon grasping and retaining the concept of variance presented in this chapter.
o Consequently, this chapter is essential to success in the remainder of the course.
o If you have any negative scores you simply first square each deviation score and then add up
the squared deviations.
o So the next step would be to calculate the variance which is basically equal to the squared
deviations, soit is the average squared distance from the mean.
o Deviations squared=Variance
~ Step Four:
o Goal: to compute a measure of the “standard” (average) distance of the scores from the mean
o Variance measures the average squared distance from the mean—not quite our goal
~ Adjust for having squared all the differences by taking the square root of the variance
Standard Deviation = Variance
o Variance (in squared distance units) is not intuitively easy to grasp despite being a measure of
average squared distance of scores from the mean.
o Consequently, it is important to emphasize the need to take the square root of the variance to
return it to the same distance unit used in the original measurement procedure.
o So no we can speak to the Standard Deviation:
o This is the square root of the variance. So in order to calculate the SD you first have to calculate
the Variance. The SD provides us with the average distance from the mean.
Figure 4.2 Calculating Variance and Standard Deviation
~ So this diagram can be used to summarize this entire process.
~ Firstly you find deviation score for each score then to get rid of the + and – signs you square each
deviation score.
~ Then we want to find the average of the squared deviation, and this is by adding up those squared
values and then divide by the number of scores this will give you the variance.
~ And finally, to find the standard deviation, you take the square root of the variance.
~ So you will square root that variance score and that will give you the SD.

Measuring Variance and Standard Deviation for a Sample


~ Goal of inferential statistics:
o Draw general conclusions about population
o Based on limited information from a sample
~ Samples differ from the population
o Samples have less variability
o Computing the variance and standard deviation in the same way as for a population would
give a biased estimate of the population values
~ This podcast will focus on measuring variance and SD for a sample.
~ By now we should know the aim of inferential statistics is to draw conclusions about a population
based on limited info from a sample.
~ Therefore it is of utmost importance to make sure any sample is truly representative of a population.
The goal is also to find patterns and results in our data.
~ The amount of variability in the data influences of easy it is to pick up those patterns.
o High variability obscures patterns that would be visible in low variability samples.
~ Samples different from the population.
~ A sample statistic is said to be biased if it overestimates or underestimates the population. However it
is relatively predictable to pick up on bias in a sample and therefore it can be corrected.
~ So sample variablility isn’t a mistake it is a measure of how much a value is different from the “true”
value.
~ So lets say as an example we find that the true weight of a population is 90kgs.
~ You take a sample and find the mean weight is 91kgs.
~ The difference here is 1kg.
~ If you sample again you might get different mean weights of 89, 87.5 etc, so the difference is a
reflection of variability in your sample.
~ Relating to another example: the speedometer on your car.
~ Maybe your speedometer consistently shows speed that is 5km per hour slower than you are actually
going, it does not mean that the speedometer is useless, its simply that maybe your speedomter
needs an adjustment.
~ The main aim of that adjustment is then to make sure that the resulting value for the sample is
accurate and unbiased.
Formulas for Sample Variance and Standard Deviation
~ Sum of squares (SS) is computed as before
~ Formula for variance has n-1 rather than N in the denominator
~ Notation uses s instead of σ

~ When you are calculating variance for samples the only difference is that the denominator has this n-1
adjustment.
~ For samples, we take the sample size minus 1. the variance for a sample is equal to the sum of
squares divided by the sample size minus 1 and in order to get the SD we just take the square root of
the variance.
~ So when receiving any sample of data and you need to calculate variance and SD for a sample you
make that n-1 adjustment.
~ So why do we do this?....
Figure 4.4 Population of Adult Heights
~ The population of adult heights forms a normal distribution.
~ We know this because when we draw a line through the middle, it will be symmetrical.
~ If you select a sample from this population, you are most likely to obtain individuals who are near
average in height.
~ As a result, the scores in the sample will be less variable (spread out) than the scores in the
population.
~ So, the reason is that samples underestimate the true variability in the population just like we spoke
of in the first slide.
~ Here you can see the entire population and if we just sample 10 or 15 of these individuals we might
me lead to believe there’s less variability inherent in the distribution than really exists.
~ So, by adding that adjustment it then reduces the denominator values and inflates the SD slightly to
correct that initial underestimate.
~ So, let’s look at some examples to calculate this in the video next
Example: Calculating Sum of Squared Deviations (SS)
(å " ! )
SS=åx2 - $
10,7,6,10,6,15 n=6
10+7+6+10+6+15=54 åx=54
10 +7 +6 +10 +6 +15 =546
2 2 2 2 2 2
åx2=546
(%&)!
SS=546- '
=546-486
=60
Example: Calculating Sample Variance
((
s2 =$)* SS=60
n=6
'+
s2 =')*
'+
=%
s2 =12
Example: Calculating Standard Deviation
S=√12 s2 =12
= 3.46
Sample Variability and Degrees of Freedom
~ Population variance
o Mean is known
o Deviations are computed from a known mean
~ Sample variance as estimate of population
o Population mean is unknown
o Using sample mean restricts variability
~ Degrees of freedom
o Number of scores in sample that are independent and free to vary
o Degrees of freedom (df) = n – 1
~ So we know that with the variance for a population the mean is known, unlike with variance for a
sample where the mean of a population is unkown.
~ Hence we measure distance from the sample mean.
~ And we know that we must first compute the sample mean before we can begin to compute
deviation.
~ But calculating the value of M places a restriction on the variability of scores in a sample.
~ This can be demonstrated in the following table.
X A sample of n=3 scores with a mean of 5

2
9
------ (What is the third score?)
~ For example you have sample of n=3 scores and compute a mean of M=5.
~ The first two scores in the sample have no restrictions and we can see that the third score is
restricted, so what do we do?
~ We see that the third score must be 4. we say 4 because the entire sample has a Mean of 5 so for 3
scores to have a mean of 5 the total must be 15.
~ So get to 15 we can see that the first two scores added together gives us 11 therefore 4 is left to make
up 15, hence we say that 4 is the restricted value.
~ So the first two scores were free to have any value but the third score was dependent on the first
two.
~ So a the first sample of n-1 scores are free to vary but the final score is restrict therefore we say that
as a result, the sample is said to have n-1 degrees of freedom.
~ The degrees of freedom determine the number of scores in a sample that are free to vary.
Chapter 5: z-Scores: Location of Scores and Standardized Distributions
Introduction & Purpose of z-scores
~ Identify and describe location of every
score in the distribution
~ Standardize an entire distribution
~ Take different distributions and make them equivalent and comparable
~ In this chapter we will discuss z-scores and the location in a distribution.
~ So, we will learn how we can take any raw score of a test or an assessment for example and convert it
into a standardized score.
~ Okay, so when we talk about z-scores we talk about turning a raw score value into a z-score or a
standardized score.
~ Why?
~ Two purposes: when we convert a raw score into a standard score, we are able to tell the exact
location of an original score of the entire sample of people that took the test.
~ Now more importantly z-scores or standardized scores allow us to directly compare the results of
scores that come from two completely different distributions that have their own mean and own
standard deviation.
~ So, by taking those two raw scores and putting it on a scale we can see which is more or less
competitive than the other.
~ One practical example you can use to introduce z-scores and standardization involves baseball.
~ How could you compare the performance of a player in 1968 to one in 2000?
~ We know that scoring was much lower in 1968, so if we simply looked at their raw batting averages,
or home runs hit, the player from 1968 would likely appear to be much worse.
~ But if we standardize their scores—by comparing them to the mean batting average (or home runs
hit) in 1968 and 2000, we now have a common metric to compare the scores: by how well they did
relative to other players in that season.
Figure 5.1 Two Exam Score Distributions

~ Two distributions of exam scores.


~ For both distributions, μ = 70, but for one distribution, σ = 3 and for the other, σ = 12.
~ The relative position of X = 76 is very different for the two distributions.
z-Scores and Location in a Distribution
~ Exact location is described by z-score
o Sign tells whether score is located above or below the mean
o Number tells distance between score and mean in standard deviation units
~ So calculating z-scores is relatively simple.
~ So z-scores can either be positive or negative and this tells us that the result is either below or above
the average.
~ The number itself tells us how many how many sd units or brackets that score lies above or below
the mean.
~ For example: If I calculate a z-score of +2.0 that means that the score is 2 full SD brackets above the
mean.
~ Alternatively a score of -2.0 tells us that the score lies 2 full SD brackets below the mean.
Figure 5.2 Relationship Between z-Scores and Locations

~ Relationship between z-score values and locations in a population distribution.


~ Okay, so here is our z-distribution, the mean of the z-distribution is 0 and the SD is 1.
~ Those are the characteristics of z curve.
~ This is important to remember.
~ Right, so we will look at computing z-scores for sample and population in the next podcast.
Computing z-Scores for Samples and Populations
~ Populations are most common context for computing z-scores
~ It is possible to compute z-scores for samples
o Indicates relative position of score in sample
o Indicates distance from sample mean
~ Sample distribution can be transformed into z-scores
o Same shape as original distribution
o Same mean M and standard deviation s
~ In today’s podcast we will touch on computing z-scores for samples and populations.
~ So populations are the most common context for computing z-scores usually we want to see where
someone falls relative to the rest of the population however we can also compute z-scores for
samples.
~ So the definition and purpose of a z-score is the same for samples and populations provided that the
mean and sample mean and sample standard deviation are used to specify a z-score location.
~ So when computing z-scores for samples and populations, the X value is transformed into a z-score
so that:
~ The sign of the z-score indicates whether the X value is + or - so above or below the mean.
~ Now lets look at the equations for calculating the z-score for samples and populations before we
actually get into the way in which to calculate these scores…
Equation for z-Score for a Population

~ Numerator is a deviation score


~ Denominator expresses deviation in standard deviation units
~ Remember that z-scores identify a specific location of a score in terms of deviations from the mean
and relative to the standard deviation.
~ Okay, here is the equation for a z-score of a population.
~ So in order to take any raw score and turn it int a z-score, we simply take that score and subtract the
mean from that value and then we divide by the SD.
~ So essentially we look and say how different is that value (X) from the population mean (u) and we
divide by the SD.
~ Look at the video in the next slide for an example on how to calculate the z-score.
Calculating z-Score for a Population
~ Remember that z-scores identify a specific location of a score in terms of deviations from the mean
and relative to the standard deviation.
~ Okay, here is the equation for a z-score of a population.
~ So in order to take any raw score and turn it int a z-score, we simply take that score and subtract the
mean from that value and then we divide by the SD.
~ So essentially we look and say how different is that value (X) from the population mean (u) and we
divide by the SD.
~ Look at the video in the next slide for an example on how to calculate the z-score.
")µ
z= s
µ=100 s=10 x=130
*,+)*++
z= *+
,+
=*+
z=3.0
µ=60 s=9 x=65
'%)'+
z= -
%
z=-
z=0.55 / 0.56
Equation for z-Score for a Sample

~ Numerator is the score - mean


~ Denominator expresses standard deviation
~ So the z-score for a sample is expressed in this formula here where each X value in a sample can
be transformed into z-scores by using this formula.
~ So we say that z-score equals raw score value which we input as the X value minus the mean
then divided by the standard deviation.
~ Here for samples the standard deviation is symbolized with the letter s.
Calculating z-Score for a Sample
~ Remember that z-scores identify a specific location of a score in terms of deviations from the mean
and relative to the standard deviation.
~ Okay, here is the equation for a z-score of a population.
~ So in order to take any raw score and turn it int a z-score, we simply take that score and subtract the
mean from that value and then we divide by the SD.
~ So essentially we look and say how different is that value (X) from the population mean (u) and we
divide by the SD.
~ Look at the video in the next slide for an example on how to calculate the z-score.
").
z=
(
M=40 x=35 s=10
,%)&+
z= *+
)%
= *+
z = -0.50
M=60 x=48 s=9
&/)'+
z= -
)*0
= -
z = -1.33
Determining a Raw Score (X) from a z-Score for Populations
X -µ
z= so
s X = µ + zs
~ Algebraically solve for X to reveal that…
~ Raw score is simply the population mean plus (or minus if z is below the mean) z multiplied by
population standard deviation
~ So although the z-score equation works well for transforming X values into z-scores, it can be a bit
awkward when you are trying to work in the opposite direction by changing z-scores back into X
values.
~ So this is done by using the equation on the right. X= mean (population) plus z score multiplied by
the standard deviation.
~ Remember the z score describes exactly where the score is located by identifying the direction and
distance from the mean.
~ So lets look at some examples in the video on the next slide…
Examples
x=µ+zs
z=2.00 µ=60 s=9
=60+18(2.00x9)
x=78
Determining a Raw Score from z-Score for Samples
~ Similarly each z-score can be transformed back into a raw score with this formula here for samples.
~ We are looking for the X value so we say X equals the Mean + the z-score multiplied by the standard
deviation.
~ Lets look at some examples in the video on the next slide.
Examples
x=M+zs
M=40 z=1.00 s=10
=40+10
x=50
Using z-Scores to Standardize a Distribution
~ Now we spoke about this in the previous podcast and we know that every X value can be
transformed into a corresponding z-score.
~ Now more specifically, if every X value is transformed into to a z-score, then the distribution of the z-
score will have these characteristics:
~ shape: the distribution of z-scores will have the same as the original distribution of scores.
~ Every X value can be transformed to a z-score
~ Characteristics of z-score transformation:
o Same shape as original distribution
o Mean of z-score distribution is always 0
o Standard deviation is always 1.00
~ A z-score distribution is called a standardized distribution
~ So for example if the original distribution is positively skewed, the z-score distribution will also be
positively skewed.
~ So as I have mentioned before, if we have a z-score of +1.00 this means that it is above the mean by
1. likewise if its -1.00 it means that it is below the mean by 1.00.
~ 2. The mean: In a z-score distribution the mean will always be zero.
~ So all positive z-scores are above the mean of zero and all negative z-scores are below the mean of
zero.
~ 3. The standard deviation: the distribution of z-scores will then always have a SD of 1.
Figure 5.5 Transforming a Population of Scores
~ An entire population of scores is transformed into z-scores.
~ The transformation does not change the shape of the population, but the mean is transformed into a
value of 0 and the standard deviation is transformed to a value of 1.
Using z-Scores for Making Comparisons
~ All z-scores are comparable to each other
~ Scores from different distributions can be converted to z-scores
~ z-scores (standardized scores) allow the direct comparison of scores from two different distributions
because they have been converted to the same scale
~ Remember we said that we can compare z-scores to each other.
~ So whether you look at Karen’s Psychology test score and compare it with her English test score, we
can directly compare the scores because they have been converted into a z-score.
Chapter 6: Probability
The Unit Normal Table
~ The proportion for only a few z-scores can be shown graphically
~ The complete listing of z-scores and proportions is provided in the unit normal table
~ Unit Normal Table is provided in Appendix B, Table B.1
~ This podcast will demonstrate how to use the Unit Normal Table.
~ So this table lists the proportions of the normal distribution for a full range of possible z-scores.
~ The table that I will show you next is just a portion of the complete table but you will be able to find
the compete table in you textbook under the appendices.
Figure 6.6 Portion of the Unit Normal Table

~ So this Unit Normal table we can see at the top of each column A, B, C, and D.
~ So remember what we learnt in the previous podcasts that transforming an x-value we use the
equation z= x-u divided by SD for populations and for samples we use z=x-m divided by s.
~ Son once you transform your raw score into the z-score you enter the z-score into the table under
column A, but this is given here already in the table.
~ Please also note that the z-score is inputted as 2 digits right at the decimal, so adhere to the standard
rounding rules.
~ The next column B represents Proportion in the body, and over here we can see an illustration of
what it looks like.
~ So if we are interested in the area or probability of proportion in a distribution below a positive z-
score, again this would be positive because it is above the mean, we would sketch the distribution like
this and we would transform the x-value into a z-score and then report the are in column B which
represents the proportion in the body which will always be greater that point 5 or 50%.
~ Here we can mirror the negative z-score
1)µ
z= s
~ So again I say a normal distribution is symmetrical right, so if we look at a negatively skewed
distribution where we could have a negative z-score so the value is below the mean.
~ Now in the next Column C, we would be interested in the area above a positive z-score.
~ So a value above a positive z-score which we may tend to see as the smaller are in the distribution.
~ So because the distribution is symmetrical we can say the same for the other side in terms of below a
negative z-score, again the proportion or probability of a score that is less than the mean.
1)2
z= (

~ Finally we have column D, which is refereed to as the proportion between the mean and z.
~ So we may be interested between the proportion and probability of score between the mean and a z-
score.
~ So here in the distribution it pertains to a positive z-score and we can consider the same, the area
between the mean and the z for a negative z-score.

~ Okay, so in the next podcast I will show you how to use the unit normal table to find proportions and
probabilities for a specific z-score.
Probability/Proportion & z-scores
~ Unit normal table lists relationships between z-score locations and proportions in a normal
distribution.
~ If you know the z-score, you can look up
the corresponding proportion.
~ If you know the proportion, you can use the table to find a specific z-score location.
~ Probability is equivalent to proportion.
~ In this podcast I will show you how to find the Proportions / probabilities for a specific z-score and Z-
Score locations that correspond to specific proportions.
~ There are a few steps to follow:
~ Firstly, to find proportions or probabilities for a specific z-score value you always:
~ Sketch the distribution and shade the are of interest.
~ Transform X values into z-scores
~ Enter the Unit Normal Table using column A and reference the appropriate proportion (Body, Tail, or
Area between Mean and z) according to the sketch drawn in the first step.
~ Remember it is vital to first sketch the distribution. If you dont you may make careless mistakes.
~ Now lets look at an example… In the next slide
~ Remember our equations for z-score…
Example
~ A sample is normally distributed with a mean of u=45 and a standard deviation of o=4.
~ What is the probability of randomly selecting a score that is greater than 43?
Refer to the Table

Z-Score locations that corresponds to specific proportions


Example
~ What z-score represents the top 10 % from the remainder of the distribution?
µ=24.3 min
s=10 min
10%
10%/100 =0.10

x=µ+s(z)
=24.3+10(1.28)
=24.3+12.8
=37.1min. ->the 10%
The Unit Normal Table to Locate the z-Score
Calculating the X value corresponding to proportions or probability
~ The probabilities given in the unit normal table will be accurate only for normally distributed scores so
the shape of the distribution should be verified before using it.
~ For normally distributed scores
o Transform the X scores (values) into z-scores
o Look up the proportions corresponding to the z-score values
Chapter 7: Probability and Samples: The Distribution of
Sample Means
Shape, Central Tendency, and Variability for the Distribution of Sample Means:
• The central limit theorem is used to specify the shape, central tendency, and variability of the
distribution of sample means.
• The standard error of M measures how well an individual sample mean represents the entire
distribution.
• So, the central limit theorem is used to specific the shape, central tendency and variability of a
distribution of sample means.
• It provides a description of the distribution that would be obtained if you selected every possible
sample mean and constructed a distribution of the sample mean.
• Then the standard error of the mean measure how well an individual sample mean represents the
entire distribution.
Central Limit Theorem
• Applies to any population with mean μ and
standard deviation σ
• Distribution of sample means approaches a normal distribution as n approaches infinity
• Distribution of sample means for samples
of size n will have a mean of μM
s
• Distribution of sample means for samples of size n will have a standard deviation = n

• For any population with mean u and standard deviation o, the distribution sample means for
sample size n will have a mean of u and standard deviation of o/ square root of n and will
approach a normal distribution as n approaches infinity.
• So essentially he value of this theorem comes from the fact that it describes the distribution of
any sample means for any population, no matter what the shape, mean or standard deviation is.
• It also states that the distribution of the sample means approaches a normal distribution very
rapidly.
• So this theory follows three steps: shape, central tendency and variability…
Shape of the Distribution of Sample Means
• The distribution of sample means is almost perfectly normal in either of two conditions
• The population from which the samples are selected is a normal distribution
or
• The number of scores (n) in each sample is relatively large—at least 30
• Speaking to the shape of the distribution sample means, it tends to be normal if there are two
specific conditions:
1. the population from which the samples are selected is a normal distribution.
2. the number of scores (n) in each sample is relatively large, around 30 or more.
• So as n gets larger, the distribution of sample means will closely approximate normal distribution.
• So essentially as n gets greater than 30, the distribution is normal regardless of the shape of the
original population.
Expected Value of M
• The mean of the distribution of sample means is μM and has a value equal to the mean of the
population of scores, μ.
• The mean of the distribution of sample means is called the expected value of M.
• M is an unbiased statistic because μM, the expected value of the distribution of sample means, is
the value of the population mean, μ.
• So, as we said in the previous slide, the sample means is centred around the mean of the
population from which samples were obtained.
• We can then go one step further and say that in actual fact, the average value of all sample means
is exactly equal to the value of the population mean.
• This mean value is called the expected value of M.
• So in laymen’s terms, a sample mean is expected be near its population mean.
• Please also note the symbols here as well.
• So M is an unbiased statistic because on average the sample statistic produces a value that is
equal to the corresponding population.
Standard Error of M
• Variability of a distribution of scores is measured by the standard deviation.
• Variability of a distribution of sample means is measured by the standard deviation of the sample
means and is called the standard error of M and written as σM.
• In journal articles or other textbooks, the standard error of M might be identified as “standard
error,” “SE,” or “SEM”.
• Now we need to go one step further and speak about the third characteristic known as Variability.
• So the value we will be working with is the standard deviation for the distribution of sample
means. This is identified with the symbol here and is called standard error of M.
• Note that in other articles or textbooks , the standard error of M is symbolised or identified using
Standard error, SE or SEM.
Definition of the Standard Error of M
• The standard error of M is the standard deviation of the distribution of sample means. When it is
large, the means are widely scattered.
• The standard error of M provides a measure of how much distance is expected on average
between M and μ.
• The standard error of M serves the same two purposes for the distribution of sample means as we
spoke about in the previous podcast.
• Firstly, the SE describes the distribution of sample means.
• It provides a measure of how much difference is expected from one sample to another.
• So we note that when the SE is small, all the sample means are close together.
• When the SE is large, the sample means are widely scattered which means there are big
differences from one sample to another.
• Secondly, SE measures how well an individual sample mean represents the entire distribution.
• It provides a measure of how much distance is reasonable to expect between an sample mean
and the overall mean of the distribution.
• We should also remember that although a sample mean is representative of the population mean,
there typically is some error between the sample and the population.
• This is where the SE measures exactly how much error is expected on average between the
sample mean and the population mean.
Standard Error Magnitude
• Law of large numbers: the larger the sample size, the more probable it is that the sample mean
will be close to the population mean.
• Population variance: The smaller the variance in the population, the more probable it is that the
sample mean will be close to the population mean.
• So here we note the law of larger numbers states that the larger the sample size the more
probable it is that the sample mean will be close to the population mean.
• In terms of population variance, the smaller the variance in the population, the more probable it is
that the sample mean will be close to the population mean.
Figure 7.3 Standard Error and Sample Size Relationship

• The relationship between standard error and sample size.


• As sample size is increased, there is less error between the sample mean and the population
mean.
• In the graph below we can see a distribution of scores in a sample and the SE of the distance
between a sample mean and the population mean.
• So, we can see here that as the sample size increased, there is less error between the sample
mean and the population mean.
A z-Score for Sample Means
• Sign tells whether the location is above (+) or below (-) the mean
• Number tells the distance between the location and the mean in standard deviation (standard
error) units
M -µ
z=
• z-formula: sM

• The z-score tells us exactly where the sample mean is located in relation to all other possible
sample means that could have been obtained.
• So a z-score identifies the location with a signed number.
• The sign tells us whether the location is above or below the mean, we should know this by now.
• So if its positive then it’s above the mean, and if its negative it is below the mean.
• The number then also tells us the distance between the location and the mean in terms of the
number of standard deviations.
• Look here at the formula for calculating the z score for sample means: so z equals the M minus
the SD divided by the SD Error.
Calculating a z-Score for a sample mean
• The distribution of sample means for n = 16.
• Samples were selected from a normal population with μ = 500 and σ = 100.
Chapter 8: Introduction to Hypothesis Testing
The Logic of Hypothesis Testing
• Hypothesis testing is one of the most used inferential procedures.
• Definition: a statistical method that uses sample data to evaluate the validity of a hypothesis
about a population.
• State hypothesis about a population
• Predict the expected characteristics of the sample based on that hypothesis
• Obtain a random sample from the population
• Compare the obtained sample data with the prediction made from the hypothesis
• If consistent, hypothesis is reasonable
• If discrepant, hypothesis is rejected
Four Steps in Hypothesis Testing
Step 1: State the hypotheses
Step 2: Set the criteria for a decision
Step 3: Collect data & compute sample statistics
Step 4: Make a decision
Step 1: State Hypotheses
• Null hypothesis (H0) states that the treatment has no effect: there is no change, difference, or
relationship.
• The independent variable has no effect on the dependent variable.
• Alternative hypothesis (H1) states that there is a change, difference, or relationship in the general
population.
Step 2: Set the Decision Criterion
• Distribution of sample outcomes is divided
– Those likely if H0 is true
– Those “very unlikely” if H0 is true
• Alpha level, or significance level, is a probability value used to define “very unlikely” outcomes.
• Critical region(s) consist of the extreme sample outcomes that are “very unlikely.”
• Boundaries of critical region(s) are determined by the probability set by the alpha level.
Figure 8.3 Note “Unlikely” Parts of Distribution of Sample Means

o The set of potential samples is divided into those that are likely to be obtained and those that are
very unlikely to be obtained if the null hypothesis is true.
Figure 8.4 Critical region(s) for α = .05

o The critical region (very unlikely outcomes) for alpha = .05.


Step 3: Collect Data (and…)
• Data always collected after hypotheses stated
• Data always collected after establishing decision criteria
• This sequence assures objectivity
• Compute a sample statistic (z-score) to show the exact position of the sample.
M -µ
z=
sM
• In words, z is the difference between the observed sample mean and the hypothesized population
mean divided by the standard error of the mean.
Step 4: Make a decision
• If sample statistic (z) is in the critical region, the null hypothesis is rejected.
• If the sample statistic (z) is not located in the critical region, the researcher fails to reject the null
hypothesis.
• M = 74, μ = 60 kg and, with n = 81 and σ = 6, the standard error for the sample mean is
= 6/9 = 0.6
= (74– 60) / 0.6 = 23.3
• With an Alpha = .05, this z-score is beyond the boundary of 1.96. The z-score is in the critical
region, and we can reject the H0
Jury Trial: Hypothesis Testing Analogy
• Trial begins with the null hypothesis “not guilty” (defendant’s innocent plea).
• Police and prosecutor gather evidence (data) relevant to the validity of the innocent plea.
• With sufficient evidence against, jury rejects null hypothesis innocence claim to conclude “guilty.”
• With insufficient evidence against, jury fails to convict, i.e., fails to reject the “not guilty” claim (but
does not conclude defendant is innocent).
Uncertainty and Errors in Hypothesis Testing
• Hypothesis testing is an inferential process
• It uses limited information from a sample to make a statistical decision, which is then used
to make a general conclusion
• Sample data used to make the statistical decision allows us to make an inference and draw
a conclusion about a population
• Errors are possible
Type I Errors
• Researcher rejects a null hypothesis that is true.
• Researcher concludes that a treatment has an effect when it has none.
• Alpha level is the probability that a test will lead to a Type I error.
Type II Errors
• Researcher fails to reject a null hypothesis that is actually false.
• Researcher has failed to detect a real treatment effect.
• Type II error probability is not easily identified.
Table 8.1

Hypothesis Testing Summary


• Step 1: State hypotheses and select alpha level
• Step 2: Locate the critical region
• Step 3: Collect data; compute the test statistic
• Step 4: Make a probability-based decision about H0: Reject H0 if the test statistic is unlikely when
H0 is true—called a “significant” or “statistically significant” result
Directional Hypothesis Tests
• The standard hypothesis testing procedure is called a two-tailed (non-directional) test because
the critical region involves both tails to determine if the treatment increases or decreases the
target behavior.
• However, sometimes the researcher has a specific prediction about the direction of the treatment.
• When a specific direction of the treatment effect can be predicted, it can be incorporated into the
hypotheses.
• In a directional (one-tailed) hypothesis test, the researcher specifies either an increase or a
decrease in the population mean because of the treatment.
Figure 8.6 Critical Region for Directional Test
One-tailed and Two-tailed Tests Compared
• A one-tailed test allows rejecting H0 with relatively small difference provided the difference is in
the predicted direction.
• A two-tailed test requires a relatively large difference regardless of the direction of the difference.
• In general, two-tailed tests should be used unless there is a strong justification for a directional
prediction.
Chapter 9: Introduction to the t Statistic
The t statistic: An alternative to z
• Sample mean (M) estimates (& approximates) population mean (μ)
• Standard error describes how much difference is reasonable to expect between M and μ.
either or

s s2
sM = sM =
n n
Reviewing the z-Score Statistic
• Use z-score statistic to quantify inferences about the population.

• Use unit normal table to find the critical region if z-scores form a normal distribution
The Problem with z-Scores
• The z-score requires more information than researchers typically have available
• Requires knowledge of the population standard deviation σ
• Researchers usually have only the sample data available
Estimated standard error
• Use s2 to estimate σ2.
• Estimated standard error:
s s2
estimated standard error = sM =
or
n n
• Estimated standard error is used as estimate of the real standard error when the value of σM is
unknown.
Introducing the t Statistic
• t statistic is an alternative to z.
• t might be considered an “approximate” z.
• Estimated standard error (sM) is used as in place of the real standard error when the value of σM is
unknown.
The t-Statistic
• The t statistic uses the estimated standard error in place of σM.
M -µ
t=
sM
• The t statistic is used to test hypotheses about an unknown population mean μ when the value of
σ is also unknown.
Degrees of Freedom
• Computation of sample variance requires computation of the sample mean first
• Only n-1 scores in a sample are independent
• Researchers call n-1 the degrees of freedom
• Degrees of freedom
• Noted as df
• df = n − 1
Figure 9.1: Distributions of the t Statistic

• Distributions of the t statistic for different values of degrees of freedom are compared to a normal
z-score distribution.
• Like the normal distribution, t distributions are bell-shaped and symmetrical and have a mean of
zero.
• However, t distributions are more variable, indicated by the flatter and more spread-out shape.
• The larger the value of df is, the more closely the t distribution approximates a normal
distribution.
The t Distribution
• Family of distributions, one for each value of degrees of freedom
• Approximates the shape of the normal distribution
• Flatter than the normal distribution
• More spread out than the normal distribution
• More variability (“fatter tails”) in t distribution
• Use Table of Values of t in place of the Unit Normal Table for hypothesis tests
The t distribution table
Figure 9.2

The t Distribution for df = 3


• The t distribution with df = 3.
• Note that 5% of the distribution is in the tail beyond t = 2.353.
• Also, 5% is in the tail beyond t = -2.353.
• Thus, a total proportion of 10% (0.10) is in the two tails beyond t = ±2.353.
Hypothesis Tests with the t Statistic
• The one-sample t test statistic (assuming the null hypothesis is true)
sample mean - population mean M - µ
t= = =0
estimated standard error sM
Figure 9.3 Basic Research Situation for t Statistic

• The basic research situation for using the t statistic hypothesis test. It is assumed that the
parameter μ is known for the population before treatment.
• The purpose of the research study is to determine whether the treatment has an effect.
• Note that the population after treatment has unknown values for the mean and the variance.
• We will use a sample to test a hypothesis about the population mean.
Hypothesis Testing: Four Steps
• State the null and alternative hypotheses and
select an alpha level
• Locate the critical region using the t distribution table and df
• Calculate the t test statistic
• Make a decision regarding H0 (null hypothesis)
Figure 9.4: Critical Region in the t Distribution for α = .05 and df = 8

• The critical region in the t distribution for α = .05 and df = 8.


Assumptions of the t Test
• Values in the sample are independent observations.
• The population sampled must be normal
• With large samples, this assumption can be violated without affecting the validity of the
hypothesis test
Influence of Sample Size and Sample Variance
• The larger the sample, the smaller the error.
• The larger the variance, the larger the error.
Chapter 10: The t Test for Two Independent Samples
Independent-Measures Design Introduction
• Most research studies compare two (or more) sets of data
• Data from two completely different, independent participant groups (an independent-
measures or between-subjects design)
• Data from the same or related participant group(s) (a within-subjects or repeated-
measures design)
• Instructors may wish to bring to students’ attention the fact that there are many terms in wide
use applied to this distinction, so they need to be prepared to learn the fundamentals of the
distinction and not just memorize some phrases.
• Computational procedures are considerably different for the two designs.
• Each design has different strengths and weaknesses.
• Consequently, only between-subjects designs are considered in this chapter; repeated-measures
designs will be reserved for discussion in Chapter 11.
Figure 10.1 Independent- Measures Research Design

• The structure of an independent-measures research study.


• Two separate samples are used to obtain information about two unknown populations or
treatment conditions.
Independent-Measures Design t Statistic
• Null hypothesis for independent-measures test
H 0 : µ1 - µ2 = 0
• Alternative hypothesis for the independent- measures test
H1 : µ1 - µ2 ¹ 0
Independent-Measures Hypothesis Test Formulas
• Basic structure of the t statistic
( M 1 - M 2 ) - ( µ1 - µ 2 )
t=
s(M - M )
1 2

• t = [(sample statistic) – (hypothesized population parameter)] divided by the estimated standard


error
• Independent-measures t-test
Estimated Standard Error
• Measure of standard or average distance between sample statistic (M1-M2) and the population
parameter
• How much difference it is reasonable to expect between two sample means if the
null hypothesis is true (equation 10.1)

s2
s22
= +
1
s( M - M 2)
1
n1 n2
Box 10.1: Variability of Difference Scores
• Why add sample measurement errors (squared deviations from mean) but subtract sample means
to calculate a difference score?
• Some instructors may want to advise students that the material in this box is designed to be
background support for understanding variability and that some students find it very insightful.
Figure 10.2 Two Population Distributions

Two population distributions.


• The scores in population I vary from 50 to 70 (a 20-point spread), and the scores in population II
range from 20 to 30 (a 10-point spread).
• If you select one score from each of these two populations, the closest two values are X1 = 50 and
X2 = 30.
• The two values that are the farthest apart are X1 = 70 and X2 = 20.
Pooled Variance
• Equation 10.1 shows standard error concept but is unbiased only if n1 = n2.
• Pooled variance (sp2) provides an unbiased basis for calculating the standard error
SS1 + SS 2
s 2p =
df1 + df 2
Estimated Standard Error for the Difference between Two Means
• The estimated standard error of M1 − M2 is then calculated using the pooled variance estimate
s 2p s 2p
s( M 1 - M 2 ) = +
n1 n2
Example
• n1 = 4; SS = 72
• n2 = 8; SS = 168
• s2p = (72 + 168) / [(4-1) + (8-1) = 240/10 = 24
!" !"
• s(M1-M2) =" " + #
= √6 + 3 = 3
Resulting formula and Degrees of Freedom
• The resulting final formula for t statistic is
( M 1 - M 2 ) - ( µ1 - µ 2 )
t=
s(M - M )
1 2

• Degrees of freedom (df) for t statistic is df for first sample + df for second sample
df = df1 + df 2 = (n1 - 1) + (n2 - 1)
Note: this term is the same as the denominator of the pooled variance
Example
Chapter 11: The t Test for Two Related Samples
Introduction to Repeated-Measures Designs
• Repeated-measures design
• Also known as within-subjects design
• Two separate scores are obtained for each individual in the sample
• Same subjects are used in both treatment conditions
• No risk of the participants in each treatment group differing significantly from each other
Example

t Statistic for a Repeated- Measures Research Design


• Structurally like the other t statistics
• Essentially the same as the single-sample t
• Based on difference scores (D) rather than raw scores (X)
• Difference score = D = X2 − X1

MD =
åD
• Mean Difference: n

The Hypotheses for Repeated-Measures Design


• H0: μD = 0
• H1: μD ≠ 0
Figure 11.1 Difference Scores for 4 People Measured Twice

• A sample of n = 4 people is selected from the population.


• Each individual is measured twice, once in treatment I and once in treatment II, and a difference
score, D, is computed for each individual.
• This sample of difference scores is intended to represent the population.
• Note that we are using a sample of difference scores to represent a population of difference
scores.
• Note that the mean for the population of difference scores is unknown.
• The null hypothesis states that there is no consistent or systematic difference between the two
treatment conditions, so the population mean difference is μD = 0.
t- Statistic for Related Samples
SS
variance of D = s 2 =
df
M D - µD
t= s2
sM D sM D =
n
Example

= 9.6

M D - µD
t=
sM D
t = (5.5-9.6) / 2
t = -2.05

Repeated-Measures Design Hypothesis Tests


• Numerator of t statistic measures actual difference between the data MD and the hypothesis μD
• Denominator measures the standard difference that is expected if H0 is true
• Same four-step process as other tests
Example
Directional Hypotheses and One-Tailed Tests
• Researchers often have specific predictions for related-samples designs.
• Null hypothesis and research hypothesis are stated directionally, e.g.
• H0: μD ≤ 0
• H1: μD > 0
• Critical region is in one tail k
Example
Related-Samples t Test Assumptions
• Observations within each treatment condition must be independent.
• Population distribution of difference scores (D values) must be normally distributed.
• This assumption is not typically a serious concern unless the sample size is small.
• With relatively large samples (n > 30) this assumption can be ignored
Chapter 14: Correlation and Regression
Introduction to Correlation
• Measures and describes the relationship between two variables
• Characteristics of relationships
• Direction (negative or positive; indicated by the sign, + or −, of the correlation coefficient)
• Form (linear is most common)
• Strength or consistency (varies from 0 to 1)
• Characteristics are all independent
• Instructors may wish to note that the correlation coefficient cannot be smaller than -1 nor larger
then +1.
• If calculation produces a number outside those boundaries, an error was made.
Figure 14.1 Scatterplot for Correlational Data

• For example, a researcher could check high school records (with permission) to obtain a measure of
each student’s academic performance, and then survey each family to obtain a measure of income.
• The resulting data could be used to determine whether there is a relationship between high school
grades and family income.
• Notice that the researcher is not manipulating any student’s grade or any family’s income, but is
simply observing what occurs naturally.
• You also should notice that a correlation requires two scores for each individual (one score from each
of the two variables).
• These scores normally are identified as X and Y.
• The pairs of scores can be listed in a table, or they can be presented graphically in a scatter plot
(Figure 14.1).
• In the scatter plot, the values for the X variable are on the horizontal axis and the Y values are listed
on the vertical axis.
• Each individual is represented by a single point in the graph so that the horizontal position
corresponds to the individual’s X value and the vertical position corresponds to the Y value.
• The value of a scatter plot is that it allows you to see any patterns or trends that exist in the data.
• The data points in Figure 14.1, for example, show a clear relationship between family income and
student grades; as income increases, grades also increase
Figure 14.2 Positive and Negative Relationships

• A recent report examined the relationship between income and weight for both men and women
(Judge & Cable, 2011).
• The results of the study are shown in Figure 14.2 and provide examples of positive and negative
relationships.
• For the men [Figure 14.2(a)], there is a positive relationship between weight and income, with income
increasing as weight increases from person to person.
• For the women [Figure 14.2(b)], the relationship is negative, with income decreasing as weight
increases.
• Although there is no definite explanation for the two relationships, the authors suggest that being
thin is seen as a sign of success and self-discipline for women.
• For men, however, a degree of extra weight is often viewed as a sign of success.
• FIGURE 14.2 Examples of positive and negative relationships.
(a) Income is positively related to weight for men.
(b) Income is negatively related to weight for women.
Figure 14.3 Different Linear Relationship Values

• FIGURE 14.3 Examples of difference values for linear correlations: (a) a perfect negative correlation, -
1.00; (b) no linear trend, 0.00; (c) a strong positive relationship, approximately +0.90; (d) a relatively
weak negative correlation, approximately -0.40.
The Pearson Correlation
• Measures the degree and the direction of the linear relationship between two variables
• Perfect linear relationship
• Every change in X has a corresponding change in Y
• Correlation will be −1.00 or +1.00
covariability of X and Y
r=
variability of X and Y separately
• By far the most common correlation is the Pearson correlation (or the Pearson product– moment
correlation), which measures the degree of linear relationship; that is, how well the data points fit
a straight line.
• The Pearson correlation for a sample is identified by the letter r.
• The corresponding correlation for the entire population is identified by the Greek letter rho (fancy
p), which is the Greek equivalent of the letter r.
• Conceptually, this correlation is computed by:
• When there is a perfect linear relationship, every change in the X variable is accompanied
by a corresponding change in the Y variable.
• In Figure 14.3(a), for example, every time the value of X increases, there is a perfectly
predictable decrease in the value of Y.
• The result is a perfect linear relationship, with X and Y always varying together.
• In this case, the coverability (X and Y together) is identical to the variability of X and Y
separately, and the formula produces a correlation with a magnitude of 1.00 or -1.00.
Sum of Products of Deviations (SP)
• Like SS (sum of squared deviations)
• Measures the amount of coverability between two variables
• SP definitional formula:
SP = å(X - M X )(Y - M Y )
• The calculation of the Pearson correlation requires one new concept: the sum of products of
deviations, or SP.
• This new value is similar to SS (the sum of squared deviations), which is used to measure
variability for a single variable.
• Now, we use SP to measure the amount of coverability between two variables.
• The value for SP can be calculated with either a definitional formula or a computational formula.
The definitional formula for the sum of products is___ where MX is the mean for the X scores and
MY is the mean for the Y scores.
• The definitional formula instructs you to perform the following sequence of operations: 1. Find the
X deviation and the Y deviation for each individual. 2. Find the product of the deviations for each
individual. 3. Add the products.
• Notice that this process literally defines the value being calculated; that is, the formula actually
computes the sum of the products of the deviations.
SP – Computational Formula
• Definitional formula emphasizes SP as the sum of two difference scores
• Computational formula results in easier calculations
• SP computational formula:

SP = å XY -
å X åY
n
• The computational formula for the sum of products of deviations is:
• Because the computational formula uses the original scores (X and Y values), it usually results in
easier calculations than those required with the definitional formula, especially if MX or MY is not
a whole number.
• However, both formulas will always produce the same value for SP.
• You may have noted that the formulas for SP are like the formulas you have learned for SS (sum
of squares).
• Specifically, the two sets of formulas have exactly the same structure, but the SS formulas use
squared values (X times X) and the SP formulas use products (X times Y).
Example
• The same set of n = 4 pairs of scores are used to calculate SP, first using the definitional formula,
and then using the computational formula.
• For the definitional formula, you need deviation scores for each of the X values and each of the Y
values. Note that the mean for the Xs is MX = 2.5 and the mean for the Ys is MY = 5. The
deviations and the products of deviations are shown in the following table:

• For the computational formula, you need the X value, the Y value, and the XY product for each
individual. Then you find the sum of the Xs, the sum of the Ys, and the sum of the XY products.
These values are as follows:

• Substituting the totals in the formula gives:


Figure 14.4 Example 14.3 Scatterplot

Scatter plot of the data from Example 14.3.


Pearson Correlation Calculation
• Ratio comparing the coverability of X and Y
(numerator) with the variability of X and Y separately (denominator)
SP
r=
SS X SSY
• As noted earlier, the Pearson correlation consists of a ratio comparing the coverability of X and Y
(the numerator) with the variability of X and Y separately (the denominator).
• In the formula for the Pearson r, we use SP to measure the coverability of X and Y.
• The variability of X is measured by computing SS for the X scores and the variability of Y is
measured by SS for the Y scores.
• With these definitions, the formula for the Pearson correlation becomes:
• Note that you multiply SS for X by SS for Y in the denominator of the Pearson formula
Example

• Before starting any calculations, it is useful to put the data in a scatter plot and make a preliminary
estimate of the correlation.
• These data have been graphed in Figure 14.4 in your textbook.
• Looking at the scatter plot, it appears that there is a very good (but not perfect) positive
correlation.
• You should expect an approximate value of r = +.8 or +.9.
• To find the Pearson correlation, we need SP, SS for X, and SS for Y.
• The calculations for each of these values, using the definitional formulas, are presented in the
table below. (Note that the mean for the X values is MX = 6 and the mean for the Y scores is MY =
4.)
• Note that the correlation accurately describes the pattern shown in Figure 14.4.
• The positive sign is consistent with the fact that the Y values increase as X increases from left to right
in the graph.
• Also, if a line were drawn through the data points, from the bottom left corner where the two axes
meet to the top right data point (X = 10 and Y = 6), the data points would be very close to the line,
indicating a very good correlation (near 1.00).
Pearson Correlation and z-Scores
• Pearson correlation formula can be expressed as a relationship of z-scores.

Sample : r =
åz z
X Y

n -1

Population : r =
åz z
X Y

N
• The Pearson correlation measures the relationship between an individual’s location in the X
distribution and his or her location in the Y distribution.
• For example, a positive correlation means that individuals who score high on X also tend to
score high on Y.
• Similarly, a negative correlation indicates that individuals with high X scores tend to have
low Y scores.
• Recall from Chapter 5 that z-scores identify the exact location of each individual score within a
distribution.
• Each X value can be transformed into a z-score, zX, using the mean and standard deviation for the
set of Xs.
• Similarly, each Y score can be transformed into zY.
• If the X and Y values are viewed as a sample, the transformation is completed using the sample
formula for z (Equation 5.3, page 124).
• If the X and Y values form a complete population, the z-scores are computed using Equation 5.1
(page 122).
• After the transformation, the formula for the Pearson correlation can be expressed entirely in
terms of z-scores.
Correlations and Outliers
• An outlier is an extremely deviant individual in the sample
• Characterized by a much larger (or smaller) score than all the others in the sample
• In a scatter plot, the point is clearly different from all the other points
• Outliers produce a disproportionately large impact on the correlation coefficient
• An outlier is an individual with X and/or Y values that are substantially different (larger or smaller)
from the values obtained for the other individuals in the data set.
• The data point of a single outlier can have a dramatic influence on the value obtained for the
correlation.
• The problem of outliers is a good reason for looking at a scatter plot instead of simply basing your
interpretation on the numerical value of the correlation.
• If you only “go by the numbers,” you might overlook the fact that one extreme data point inflated
the size of the correlation.
Figure 14.7 Outlier Influences Size of Correlation

• FIGURE 14.7 A demonstration of how one extreme data point (an outlier) can influence the value of a
correlation.
Correlation and Causation
• Correlation describes a relationship but does not demonstrate causation.
• Establishing causation requires an experiment in which one variable is manipulated and others
carefully controlled.
• Example 14.4 (and Figure 14.5) demonstrates the fallacy of attributing causation after observing a
correlation.
Figure 14.5 Correlation: Churches and Serious Crimes

• Hypothetical data showing the logical relationship between the number of churches and the number
of serious crimes for a sample of U.S. cities.
14.4 Hypothesis Tests with the Pearson Correlation
• Pearson correlation is usually computed for sample data but is also used to test hypotheses about
the relationship in the population.
• Population correlation is shown by Greek letter rho (ρ).
• Non-directional: H0: ρ = 0 and H1: ρ ≠ 0
Directional: H0: ρ ≤ 0 and H1: ρ > 0 or Directional: H0: ρ ≥ 0 and H1: ρ < 0
• The Pearson correlation is generally computed for sample data. As with most sample statis- tics,
however, a sample correlation is often used to answer questions about the corresponding
population correlation.
• For example, a psychologist would like to know whether there is a relationship between IQ and
creativity.
• This is a general question concerning a population.
• To answer the question, a sample would be selected, and the sample data would be used to
compute the correlation value.
• You should recognize this process as an example of inferential statistics: using samples to draw
inferences about populations.
• In the past, we have been concerned primarily with using sample means as the basis for answering
questions about population means.
• In this section, we examine the procedures for using a sample correlation as the basis for testing
hypotheses about the corresponding population correlation.
• The basic question for this hypothesis test is whether a correlation exists in the population.
• The null hypothesis is “No. There is no correlation in the population” or “The population
correlation is zero.”
• The alternative hypothesis is “Yes. There is a real, nonzero correlation in the population.”
Because the population correlation is traditionally represented by P (the Greek letter rho),
these hypotheses would be stated in symbols as
• When there is a specific prediction about the direction of the correlation, it is possible to do a
directional, or one-tailed test.
• For example, if a researcher is predicting a positive relationship, the hypotheses would be
• The correlation from the sample data is used to evaluate the hypotheses.
• For the regular, nondirectional test, a sample correlation near zero provides support for
H0 and a sample value far from zero tends to refute H0.
• For a directional test, a positive value for the sample correlation would tend to refute a
null hypothesis stating that the population correlation is not positive.
Figure 14.9 Correlation in Sample vs. Population

• Although sample correlations are used to test hypotheses about population correlations, you
should keep in mind that samples are not expected to be identical to the populations from which
they come; there will be some discrepancy (sampling error) between a sample statistic and the
corresponding population parameter.
• Specifically, you should always expect some error between a sample correlation and the
population correlation it represents.
• One implication of this fact is that even when there is no correlation in the population (r = 0), you
are still likely to obtain a nonzero value for the sample correlation.
• This is particularly true for small samples.
• Figure 14.9 illustrates how a small sample from a population with a near-zero correlation could
result in a correlation that deviates from zero.
• The coloured dots in the figure represent the entire population and the three circled dots
represent a random sample.
• Note that the three sample points show a relatively good, positive correlation even through there
is no linear trend (r = 0) for the population.
• When you obtain a nonzero correlation for a sample, the purpose of the hypothesis test is to
decide between the following two interpretations:
1. There is no correlation in the population (r = 0) and the sample value is the result of sampling
error.
Remember, a sample is not expected to be identical to the population.
There always is some error between a sample statistic and the corresponding population
parameter.
This is the situation specified by H0.
2. The nonzero sample correlation accurately represents a real, nonzero correlation in the
population.
This is the alternative stated in H1.
• The correlation from the sample will help to determine which of these two interpretations is more
likely.
• A sample correlation near zero supports the conclusion that the population correlation is also zero.
• A sample correlation that is substantially different from zero supports the conclusion that there is a
real, nonzero correlation in the population.
Correlation Hypothesis Test
• Sample correlation r used to test population ρ
• Degrees of freedom (df) = n − 2
• Hypothesis test can be computed using
either t or F; only t shown in this chapter
• Use t table to find critical value with df = n − 2
r-r
t=
(1 - r 2 )
(n - 2)
• The hypothesis test evaluating the significance of a correlation can be conducted using either a t
statistic or an F-ratio.
• The F-ratio is discussed later (pages 460–462) and we focus on the t statistic here.
• The t statistic for a correlation has the same general structure as t statistics introduced in
Chapters 9, 10, and 11. (sample statistic – population parameter / standard error)
• In this case, the sample statistic is the sample correlation (r) and the corresponding parameter is
the population correlation (P).
• The null hypothesis specifies that the population correlation is r = 0.
• The final part of the equation is the standard error, which is determined by
• Thus, the complete t statistic is FORMULA
• Degrees of Freedom:
• The t statistic has degrees of freedom defined by df = n - 2. An intuitive explanation for
this value is that a sample with only n = 2 data points has no degrees of freedom.
• Specifically, if there are only two points, they will fit perfectly on a straight line, and the
sample produces a perfect correlation of r = +1.00 or r = -1.00.
• Because the first two points always produce a perfect correlation, the sample correlation is
free to vary only when the data set contains more than two points.
• Thus, df = n - 2.
Example

Alternatives to the t statistic


• Instead of computing a t statistic for the hypothesis test, you can simply compare the sample
correlation with the list of critical values in Table B.6 in Appendix B.
• To use the table, you need to know the sample size (n) and the alpha level.
• In the previous examples, we used a sample of n = 30, a correlation of r = 0.35, and an alpha level
of .05.
• In the table, you locate df = n - 2 = 28 in the left-hand column and the value .05 for either one
tail or two tails across the top of the table.
• For df = 28 and alpha = .05 for a two-tailed test, the table shows a critical value of 0.361.
• Because our sample correlation is not greater than this critical value, we fail to reject the null
hypothesis (as in Example 14.6).
• For a one-tailed test, the table lists a critical value of 0.306. This time, our sample correlation is
greater than the critical value so we reject the null hypothesis and conclude that the correlation is
significantly greater than zero (as in Example 14.7).
• As with most hypothesis tests, if other factors are held constant, the likelihood of finding a
significant correlation increases as the sample size increases.
• For example, a sample correlation of r = 0.50 produces a nonsignificant t(8) = 1.63 for a sample of
n = 10, but the same correlation produces a significant t(18) = 2.45 if the sample size is increased
to n = 20.

You might also like