Professional Documents
Culture Documents
Chapter 1 - 14
Chapter 1 - 14
~ This material is arguably in the “Top Ten Most Important” concepts the students will encounter in the
study of statistics and may merit identifying it as such.
~ The word nominal means having to do with names or labels.
o It involves classifying individuals or events into categories.
o They all have different names but are not related in any way.
o For example: if you were measuring academic majors for a group of university students, those
majors would be classified according to categories such as: psychology, sociology, business
science, biology etc. so each student in that group would be classified in a category.
~ An ordinal scale consists of categories which are organised in a sequence.
o So, measurements are ranked in class.
o Sizes etc.
o For example: in a psychology module, one could rank students according to who came first,
second, third or fourth in the course in terms of academic achievement. The fact that
categories form an ordered sequence means that there is a relationship between categories.
~ An interval scale consists of ordered categories that are all intervals of exactly the same size.
o So, it is characterized by equal intervals between scale units.
o Basically, the difference between two values are meaningful.
o For example: Person A scored 65% in her psychology 253 exam and person B scored 80% in
her psychology 253 exam. We can also then say that person B scored 15% higher than person
A and we can go further and say that it is equal to Person C scoring 70% and Person D scoring
85% (the difference between both sets of scores is equal). When thinking of this in terms of a
scale, the zero point of the scale is arbitrary, in other words it has no zero point.
~ The ratio scale is basically an interval scale with and added characteristic.
o It has an absolute zero which means that a score of zero indicates none of the variable being
measured.
o For example: number of correct answers on a student’s psychology 253 exam. It can either be
any number of correct answers as well as zero correct answers.
Three Data Structures, Research Methods, and Statistics
Data Structure I:
o Descriptive research (individual variables)
o One (or more) variables measured per individual
o “Statistics” describe the observed variable
o May use category and/or numerical variables
~ Data structure 1: One or More Separate Variables Measured for Each Individual:
~ Descriptive Research:
~ This involves measuring one or more separate variables for each person or participant.
~ The intention here is to simply describe as the title says.
~ Think of this example
Individual Number of hours Number of hours Number of hour
exercise in a day sleeping in a day studying in a day
A 2 6 4
B 1 7 3
C 3 4 5
D 4 8 2
~ So in this table we can see that it speaks to how many hours an individual exercises, sleeps and
studies in each day.
~ Here it is different for each person, and we are simply describing the variables by saying that
Person A exercises for 2 hours a day, sleeps for 6 hours and studies for 4 hours each day, and so
on with the rest of the individuals described here.
Relationships between variables
~ Is very important.
o Two (or more) variables observed and measured
o One of two possible data structures used to determine what type of relationship exists
~ Most research aims to examine whether there is a relationship between variables. For example:
~ Is there a relationship between number of hours spent studying a day and the results on a test?
~ Is there a relationship between number of hours spent studying for psychology 253 exam and the
results of the exam?
~ To establish whether there is a relationship – we first have to make observations of the variables.
Data Structure II:
o The correlational method
o One group of participants
o Measurement of two variables for each participant
o Goal is to describe type and magnitude of the relationship
o Patterns in the data reveal relationships
o Non-experimental method of study
~ Data Structure 2 involves the correlational method.
~ So, this is observing one group with two variables being measured.
~ So we just spoke about examining the relationship between variables and this can be done using two
variables.
~ Simply put, we can measure two variables for each individual.
~ You have examples in your textbook but I want to focus on another example here:
~ In the correlational method two different variables are observed to determine whether there is a
relationship between them.
~ Example
STUDENT NUMBER OF HOURS ACADEMIC
SPENT ON SOCIAL PERFORMANCE
MEDIA (Results on a test)
A 5 50%
B 6 40%
C 4 60%
D 2 70%
o This table shows information on 4 students in relation to the amount of time they spend on
social media and the results they score on a test.
o So, you can see that person C spends 4 hours on social media and scored 60% on her test
o Whereas person B spends 6 hours on social media and scored 40% on his test…
o draw scatterplot with whiteboard function.
Figure 1.5 Data structures for studies evaluating the relationship between variables
o One of two data structures for studies evaluating the relationship between variables.
o Note that there are two separate measurements for each individual (Facebook time and academic
performance).
o The same scores are shown in table (a) and graph (b).
o The second data structure for studies evaluating the relationship between variables.
o Note that one variable is used to define the groups and the second variable is measured to
obtain scores within each group.
o So here you are assessing two variables, type of video game (violent and non-violent) and the
second variable - aggressive behavior is measured to obtain the scores as seen in this table
above.
Experimental Method
~ The experimental method is systematic and scientific approach in which the research manipulates one
or more variables.
~ So, one variable is manipulated and the other is controlled.
o For example: administering a diet supplement to one group and the other group receives a
placebo; there weight-loss is being measured.
~ Goal of experimental method
o To demonstrate a cause-and-effect relationship
o This requires manipulation by the researcher changing its value from one level to another.
~ Manipulation
o The level of one variable is determined by the experimenter
o So relating back to the previous example about violent video games and non-violent video
games.
o The research then manipulates by giving one group of boys a violent video game to play and
the other group non-violent video game.
o The second variable as mentioned is the variable that is measured and in this example it is
whether the manipulation had an effect or not.
~ Control rules out influence of other variables
o Participant variables
o Environmental variables
o Regarding control, this is where the researcher makes sure that no external factors influence
the relationship.
o So in the example of the diet supplement where one group receives the diet pill and the other
group receives the ‘fake’ diet pill (placebo), the researcher then makes sure that no extraneous
factors influence the relationship by ensuring maybe the both groups do the same exercise
routine and eat the same foods etc.
Experimental Method: Control
~ Within the experimental method, there can be controlled and experimental conditions.
~ Individuals in a controlled condition do not receive the experimental treatment (for example, the
group of boys who played the on-violent video games or the one group of people who received the
placebo diet pill.
~ Methods of control
o Random assignment of subjects
o Matching of subjects
o Holding level of some potentially influential variables constant
~ Control condition
o Individuals do not receive the experimental treatment
o They either receive no treatment or they receive a neutral, placebo treatment
o Purpose: to provide a baseline for comparison with the experimental condition
o The purpose of such a condition is to provide a baseline for comparison.
~ Experimental condition
o Individuals do receive the experimental treatment
o Individuals in the experimental conditions do receive the experimental treatment (i.e the
violent video games or the diet pill)
Independent/Dependent Variables
~ Independent variable is the variable manipulated by the researcher
o Independent because no other variable in the study influences its value
~ Dependent variable is the one observed to assess the effect of treatment
o Dependent because its value is thought to depend on the value of the independent variable
~ So in our example the independent variable would be the amount of violence (so the violent vs non-
violent video games) and the dependent variable is the level of aggressive behaviour.
Non-Experimental Methods
~ Non-equivalent groups
o Researcher compares groups
o Researcher cannot control who goes into which group
~ Pre-test / Post-test
o Individuals measured at two points in time
o Researcher cannot control influence of the passage of time
~ Independent variable is quasi-independent
~ There are also non-experimental methods.
~ The first being no-equivalent groups:
o Any study that follows a scientific requirement such as the previous examples are known as
experimental studies.
o And as such there are also designs that are no-experimental but still examine the relationships
between variables.
~ A pre-test and post-test study uses time such a s before and after to create groups of scores. Lets
see the examples…
Two examples of non-experimental studies: Figure 1.7
~ Two examples of non-experimental studies that involve comparing two groups of scores.
~ In (a), a participant variable (gender) is used to create groups, and then the dependent variable
(verbal score) is measured in each group.
~ In (b), time is the variable used to define the two groups, and the dependent variable (depression)
is measured at each of the two times.
Chapter 2: Frequency Distributions
Frequency Distributions
~ A frequency distribution is
o An organized tabulation
o Showing the number of individuals located in each category on the scale of measurement
~ Terminology associated with frequency distributions is one of the least “standardized” across
disciplines and texts students might encounter.
~ Instructors may wish to emphasize the importance of being precise with the terms provided by the
text authors, but also be aware that terms may differ in other texts or courses.
~ Can be either a table or a graph
~ Always shows:
o The categories that make up the scale
o The frequency, or number of individuals, in each category
~ I am not going to go into any definitions as you as the student needs to make sure you go through
every concept and make sure you become familiar with the terminology used in statistics.
~ This is crucial…
o So we know that a frequency distribution is basically data organized in a table that can explain
something…
o It can also be distributed onto a graph like we saw in the previous podcast on data structures.
o But here you will learn the statistical input of data in tables or graphs. Like I said before
practice makes perfect, so go through your textbook and practice using the Learning Check
questions.
o So you can know more or less how you may be examined or what a certain questions asks of
you.
o In any frequency distribution it indicates the categories which make up the scale as well as the
frequency or number of individuals in each category.
Example
X f
8 9 8 7 10 9 6 4 9 8
7 8 10 9 8 6 9 7 8 8 10 2
9 5
N= (number of scores 8 7
20) 7 3
X= score
6 2
f= number of times
5 0
scores occurred.
4 1
ΣX = N
X f
5+4+4+3+3+3+2+ 5 1
2+2+1= 29 4 2
3 3
ΣX = 29 2 3
1 1
ΣX2= 97
o see this table here… we have 20 scores but we can see that some scores present more than
once so we can group it as follows…
o we can see that N= sum, X=score and f=frequency (number of times score occurred). Again
here you must basically memorize these symbols because they are used throughout stats.
§ So these set of scores are organised in the table under the column X (excluding the
scores that occur more than once) and remember it must always be tabulated in order
from highest to lowest.
§ Once you organise your scores from highest to lowest you look at each number and
see how many times a score of 10 occurred, a score of 9 occurred and so on…
§ Here you actually count how many timesa score of 10 occurred etc..
§ You can now see clearly the scores in the table. 2 people scored the highest score
which is 10 and you can also see that no one scored 5 but it is included in the table.
§ Remember the 4 different scales that we spoke about previously, with an ordinal,
interval and ratio scale the categories are listed in order rom highest to lowest.
Frequency Distribution Tables
~ Structure of frequency distribution table
o Categories in a column (often ordered from highest to lowest but could be reversed)
o Frequency count next to category
~ Σf = N
~ To compute ΣX from a table
o Convert table back to original scores or
o Compute ΣfX
~ As mentioned, the categories are ordered from highest to lowest and they are displayed under the
column X and the number of times that score appeared is found in the column f.
~ But now we want to know the total numbers of scores in the distribution and this is denoted by the
symbol above.
o to calculate the sum of scores in the distribution you need to look at both columns.
o Let’s look at the next example…
Calculating the scores in a Frequency Distribution
Consider this table…
X f ΣX
5 1 5
4 2 8
3 3 9
2 3 6
1 1 1
ΣX = 29
~ You see a set of scores and how many times the score appeared which is the frequency.
~ So to calculate the sum of the scores we do the following:
o Remember I said in the previous slide you need to consider both columns when calculating
the sum…
~ You will write out each score including the amount of times each score appeared:
~ 5+4+4+3+3+3+2+2+2+1= 29
~ To get the sum on the frequency squared you square each score and add the squared values = 97
This you do on a calculator…
~ Another way to get the sum of X and input in in a table is as follows…
~ Let’s input into the table together…
Proportions and Percentages
Proportions
• Measures the fraction of the total group that is associated with each score
f
proportion = p =
N
• Called relative frequencies because they describe the frequency ( f ) in relation to the total
number (N)
Percentages
• Expresses relative frequency out of 100
f
percentage = p(100) = (100)
N
• Can be included as a separate column in a frequency distribution table
~ The ability to quickly and comfortably convert between fractions (proportions), decimal fractions
(relative frequency), and percentages is fundamental to success in this course.
~ Some students struggle with reconciling the fact that although these are three distinct metrics, they
all point to the same “deep” meaning.
~ There are other measures that describe the distribution of scores which we can also interpret.
~ These are Proportions and Percentages.
§ Proportion measures the fraction of the total group associated with each score so for
example we see in the previous table that two people scored 4 so we can say that 2
out of 10 people had 4, this is then demonstrated in this form above.
~ Researchers also use percentages to describe a distribution, and this can be done by first finding the
proportion and then multiplying that by 100.
§ Use whiteboard to show proportion and percentage...
Example 2.4: Frequency, Proportion and Percent
µ=å
X
N
~ Sample:
M=
åX
n
~ Instructors my wish to have students compare and contrast these two formulas.
~ The mean is basically the average…
~ To calculate the mean we all all the scores and divide by the number of scores in the data.
~ The formula to calculate the mean is shown here.
~ There are two different ways to calculate the mean firstly in terms of the population and secondly for
a sample.
~ Here you can spot the slight differences between the two formulas.
~ For the formula to calculate the mean for the population it is as follows:
~ So this symbol represents the population mean and it is calculated by adding all scores and dividing
by the number of scores. So the sum of X over N.
~ For the sample mean the symbol is M=the sum of x over the number of scores…
~ So if you see a mean being represented by and M then you are dealing with a sample mean.
~ If it it being represented by the U symbol, then you are dealing with a population mean.
~ Lets looks at an example:
Example:
~ Calculating the population mean:
5, 6, 7, 3, 4
~ N= 5 scores
~ Add up all scores first: 5+6+7+3+4= 25
~ The mean is: 25
o So remember we first count how many scores there are, we see that there are 5 scores so N=5
o Then we add up all the scores which gives us 25
o Now to calculate the population mean we say pop mean = 25 divided by 5 =5 therefore our
population mean is 5.
~ Calculating the sample mean:
3, 6, 5, 3, 4
~ N= 5 scores
~ Add up all scores first: 3+6+5+3+4= 21
~ The mean is:
~ Gym attendance per week
The Weighted Mean
~ Combine two sets of scores
~ Three steps:
o Determine the combined sum of all the scores
o Determine the combined number of scores
o Divide the sum of scores by the total number
of scores
Overall Mean =
M=
åX +åX 1 2
n1 + n2
~ So what if we want to now know the mean of two sets of scores?
~ This is where we use the weighted mean formula.
~ So this formula is the Overall Mean equals the sum of scores for group 1 plus the sum of scores for
group 2 divided by the number of scores for group one and the number of scores for group two.
~ Now let’s see this in an example:
Example:
~ Hours of studying for a test per day:
~ Group 1: 3, 4, 5, 2, 6
~ Group 2: 2, 5, 3, 1, 5
~ So here we have a set of scores which indicate the number of hours a day two groups of 5 people
each, study for a test.
~ So we want to find out what the average or mean hours both groups studied for a test.
~ Group 1 hours are 3,4,5,2,6
~ Group 2 hours are 2,5,3,1,5
~ So let’s follow the steps we just spoke about for calculating the weighted mean:
~ Firstly lets add the total scores for each group: Group 1: 20 so sum of x =20
~ Group 2: 16 so the sum of x=16
~ And we know that there are a total of 5 scores per group so now lets write this out…
Computing the Mean from a Frequency Distribution Table
Quiz Score (X) f fX
10 1 10
9 2 18
8 4 32
7 0 0
6 1 6
Total n = Σf = 8 ΣfX = 66
M = ΣX / n = 66/8 = 8.25
Checkers 32
N=100 scores
Mode= 32
Woolworths 18
Pick ‘n Pay 28
Shoprite 22
~ So a basic example is this set of scores, you can see that 6 appears the most (4 times) hence 6 is the
mode.
~ Lets take another example by looking at the table above, you can see the different stores.
~ So for example a sample of 100 students on campus were asked which store they shop at for their
groceries, the most frequent answer is Checkers as 32 out of 100 students said they shopped at
Checkers.
~ It is also important to note that sometimes there can be two modes in a frequency.
~ Bimodial: is a distribution with 2 modes and a distribution with more than 2 modes is called
multimodial.
Chapter 4: Variability
Defining Variance and Standard Deviation
~ Most common and most important measure
of variability is the standard deviation
o A measure of the standard, or average, distance from the mean
o Describes whether the scores are clustered closely around the mean or are widely scattered
~ Calculation differs for population and samples
~ Variance is a necessary companion concept to standard deviation but not the same concept
o The twin concepts of variance and standard deviation are among the most challenging
concepts in a basic statistics course to communicate and to learn.
o Instructors will almost certainly want to invest special care in the preparation of materials to
help communicate these very difficult concepts.
o So in today’s podcast we wil focus on defining variance and standard deviation.
o The standard deviation is the standard or average distance ffrom the mean, its going to give us
a description of whether scores are close to or cluster together on the mean or widely
scattered out.
o Now we want to note that here too lik when calculating the mean there were slightly different
formulae, so here to calculate the SD of the population and samples are different too.
o Please ensure to familiarize yourself with the formulae in your text book as it also clearly
differentiates between sample and population formulae.
o Now Variance is a somewhat companion concept of SD. BUT it is not the same concept.
o So lets go through these formulae.
~ Step One: Determine the deviation
~ Deviation is distance from the mean
Deviation score = X − μ
~ Step Two: Find a “sum of deviations” to use as a basis of finding an “average deviation”
o Two problems
§ Deviations sum to 0 (because M is balance point)
§ If sum always 0, “Mean Deviation” will always be 0.
o Need a new strategy!
~ Having students try to come up with an intuitive method for developing a measure of variability
based on deviation scores is a great way to get them thinking about what a dead end strategy
averaging deviations is (because, of course, the average of deviations from the mean must be 0).
~ Several teams working on it in a classroom exercise often results in a valuable insight about the issue
(averaging absolute value of deviations) and might produce the one we use—squaring the deviations
to eliminate the negative values.
~ When we calculate the SD we basically asking how different the scores are from one another. So
deviation as mention means distance.
~ So for example: think of a psy 253 exam where the average grade is 70% but your score is 80%, so
essentially your score deviates from th mean by positive 10 points and likewise if you score below 80
you areminus 10 points from the mean.
~ So we can calculate for everyone in the class and calculate the deviation score for each person and at
the end we can calculate the average score for all deviations in the class.
~ This is essentially what SD is. It tells us the average distance of the sample from the pop mean.
Example
Finding the deviation score:
Deviation =x-µ µ=50 x=53
=53-50
Deviation score=3
Deviation =x-µ µ=50 x=45
=45-50
=-5
To get rid of the (-) you have to Ö (square) each and every score and then calculate the variance.
~ Step Two Revised: Remove negative deviations
o First square each deviation score
o Then sum the squared deviations (SS)
~ Step Three: Average the squared deviations
o Mean squared deviation is known as “variance”
o Variability is now measured in squared units
Population variance equals mean (average)
squared deviation (distance) of the scores
from the population mean
o The concept of sum of squared deviations (SS) is absolutely vital to efficient understanding of
the statistical tests presented in the remainder of the text.
o The authors have reduced the computational complexity and the cognitive load required of
students—contingent upon grasping and retaining the concept of SS presented in this chapter.
o The authors also lay the foundation for efficiently learning the fundamentals of ANOVA—
contingent upon grasping and retaining the concept of variance presented in this chapter.
o Consequently, this chapter is essential to success in the remainder of the course.
o If you have any negative scores you simply first square each deviation score and then add up
the squared deviations.
o So the next step would be to calculate the variance which is basically equal to the squared
deviations, soit is the average squared distance from the mean.
o Deviations squared=Variance
~ Step Four:
o Goal: to compute a measure of the “standard” (average) distance of the scores from the mean
o Variance measures the average squared distance from the mean—not quite our goal
~ Adjust for having squared all the differences by taking the square root of the variance
Standard Deviation = Variance
o Variance (in squared distance units) is not intuitively easy to grasp despite being a measure of
average squared distance of scores from the mean.
o Consequently, it is important to emphasize the need to take the square root of the variance to
return it to the same distance unit used in the original measurement procedure.
o So no we can speak to the Standard Deviation:
o This is the square root of the variance. So in order to calculate the SD you first have to calculate
the Variance. The SD provides us with the average distance from the mean.
Figure 4.2 Calculating Variance and Standard Deviation
~ So this diagram can be used to summarize this entire process.
~ Firstly you find deviation score for each score then to get rid of the + and – signs you square each
deviation score.
~ Then we want to find the average of the squared deviation, and this is by adding up those squared
values and then divide by the number of scores this will give you the variance.
~ And finally, to find the standard deviation, you take the square root of the variance.
~ So you will square root that variance score and that will give you the SD.
~ When you are calculating variance for samples the only difference is that the denominator has this n-1
adjustment.
~ For samples, we take the sample size minus 1. the variance for a sample is equal to the sum of
squares divided by the sample size minus 1 and in order to get the SD we just take the square root of
the variance.
~ So when receiving any sample of data and you need to calculate variance and SD for a sample you
make that n-1 adjustment.
~ So why do we do this?....
Figure 4.4 Population of Adult Heights
~ The population of adult heights forms a normal distribution.
~ We know this because when we draw a line through the middle, it will be symmetrical.
~ If you select a sample from this population, you are most likely to obtain individuals who are near
average in height.
~ As a result, the scores in the sample will be less variable (spread out) than the scores in the
population.
~ So, the reason is that samples underestimate the true variability in the population just like we spoke
of in the first slide.
~ Here you can see the entire population and if we just sample 10 or 15 of these individuals we might
me lead to believe there’s less variability inherent in the distribution than really exists.
~ So, by adding that adjustment it then reduces the denominator values and inflates the SD slightly to
correct that initial underestimate.
~ So, let’s look at some examples to calculate this in the video next
Example: Calculating Sum of Squared Deviations (SS)
(å " ! )
SS=åx2 - $
10,7,6,10,6,15 n=6
10+7+6+10+6+15=54 åx=54
10 +7 +6 +10 +6 +15 =546
2 2 2 2 2 2
åx2=546
(%&)!
SS=546- '
=546-486
=60
Example: Calculating Sample Variance
((
s2 =$)* SS=60
n=6
'+
s2 =')*
'+
=%
s2 =12
Example: Calculating Standard Deviation
S=√12 s2 =12
= 3.46
Sample Variability and Degrees of Freedom
~ Population variance
o Mean is known
o Deviations are computed from a known mean
~ Sample variance as estimate of population
o Population mean is unknown
o Using sample mean restricts variability
~ Degrees of freedom
o Number of scores in sample that are independent and free to vary
o Degrees of freedom (df) = n – 1
~ So we know that with the variance for a population the mean is known, unlike with variance for a
sample where the mean of a population is unkown.
~ Hence we measure distance from the sample mean.
~ And we know that we must first compute the sample mean before we can begin to compute
deviation.
~ But calculating the value of M places a restriction on the variability of scores in a sample.
~ This can be demonstrated in the following table.
X A sample of n=3 scores with a mean of 5
2
9
------ (What is the third score?)
~ For example you have sample of n=3 scores and compute a mean of M=5.
~ The first two scores in the sample have no restrictions and we can see that the third score is
restricted, so what do we do?
~ We see that the third score must be 4. we say 4 because the entire sample has a Mean of 5 so for 3
scores to have a mean of 5 the total must be 15.
~ So get to 15 we can see that the first two scores added together gives us 11 therefore 4 is left to make
up 15, hence we say that 4 is the restricted value.
~ So the first two scores were free to have any value but the third score was dependent on the first
two.
~ So a the first sample of n-1 scores are free to vary but the final score is restrict therefore we say that
as a result, the sample is said to have n-1 degrees of freedom.
~ The degrees of freedom determine the number of scores in a sample that are free to vary.
Chapter 5: z-Scores: Location of Scores and Standardized Distributions
Introduction & Purpose of z-scores
~ Identify and describe location of every
score in the distribution
~ Standardize an entire distribution
~ Take different distributions and make them equivalent and comparable
~ In this chapter we will discuss z-scores and the location in a distribution.
~ So, we will learn how we can take any raw score of a test or an assessment for example and convert it
into a standardized score.
~ Okay, so when we talk about z-scores we talk about turning a raw score value into a z-score or a
standardized score.
~ Why?
~ Two purposes: when we convert a raw score into a standard score, we are able to tell the exact
location of an original score of the entire sample of people that took the test.
~ Now more importantly z-scores or standardized scores allow us to directly compare the results of
scores that come from two completely different distributions that have their own mean and own
standard deviation.
~ So, by taking those two raw scores and putting it on a scale we can see which is more or less
competitive than the other.
~ One practical example you can use to introduce z-scores and standardization involves baseball.
~ How could you compare the performance of a player in 1968 to one in 2000?
~ We know that scoring was much lower in 1968, so if we simply looked at their raw batting averages,
or home runs hit, the player from 1968 would likely appear to be much worse.
~ But if we standardize their scores—by comparing them to the mean batting average (or home runs
hit) in 1968 and 2000, we now have a common metric to compare the scores: by how well they did
relative to other players in that season.
Figure 5.1 Two Exam Score Distributions
~ So this Unit Normal table we can see at the top of each column A, B, C, and D.
~ So remember what we learnt in the previous podcasts that transforming an x-value we use the
equation z= x-u divided by SD for populations and for samples we use z=x-m divided by s.
~ Son once you transform your raw score into the z-score you enter the z-score into the table under
column A, but this is given here already in the table.
~ Please also note that the z-score is inputted as 2 digits right at the decimal, so adhere to the standard
rounding rules.
~ The next column B represents Proportion in the body, and over here we can see an illustration of
what it looks like.
~ So if we are interested in the area or probability of proportion in a distribution below a positive z-
score, again this would be positive because it is above the mean, we would sketch the distribution like
this and we would transform the x-value into a z-score and then report the are in column B which
represents the proportion in the body which will always be greater that point 5 or 50%.
~ Here we can mirror the negative z-score
1)µ
z= s
~ So again I say a normal distribution is symmetrical right, so if we look at a negatively skewed
distribution where we could have a negative z-score so the value is below the mean.
~ Now in the next Column C, we would be interested in the area above a positive z-score.
~ So a value above a positive z-score which we may tend to see as the smaller are in the distribution.
~ So because the distribution is symmetrical we can say the same for the other side in terms of below a
negative z-score, again the proportion or probability of a score that is less than the mean.
1)2
z= (
~ Finally we have column D, which is refereed to as the proportion between the mean and z.
~ So we may be interested between the proportion and probability of score between the mean and a z-
score.
~ So here in the distribution it pertains to a positive z-score and we can consider the same, the area
between the mean and the z for a negative z-score.
~ Okay, so in the next podcast I will show you how to use the unit normal table to find proportions and
probabilities for a specific z-score.
Probability/Proportion & z-scores
~ Unit normal table lists relationships between z-score locations and proportions in a normal
distribution.
~ If you know the z-score, you can look up
the corresponding proportion.
~ If you know the proportion, you can use the table to find a specific z-score location.
~ Probability is equivalent to proportion.
~ In this podcast I will show you how to find the Proportions / probabilities for a specific z-score and Z-
Score locations that correspond to specific proportions.
~ There are a few steps to follow:
~ Firstly, to find proportions or probabilities for a specific z-score value you always:
~ Sketch the distribution and shade the are of interest.
~ Transform X values into z-scores
~ Enter the Unit Normal Table using column A and reference the appropriate proportion (Body, Tail, or
Area between Mean and z) according to the sketch drawn in the first step.
~ Remember it is vital to first sketch the distribution. If you dont you may make careless mistakes.
~ Now lets look at an example… In the next slide
~ Remember our equations for z-score…
Example
~ A sample is normally distributed with a mean of u=45 and a standard deviation of o=4.
~ What is the probability of randomly selecting a score that is greater than 43?
Refer to the Table
x=µ+s(z)
=24.3+10(1.28)
=24.3+12.8
=37.1min. ->the 10%
The Unit Normal Table to Locate the z-Score
Calculating the X value corresponding to proportions or probability
~ The probabilities given in the unit normal table will be accurate only for normally distributed scores so
the shape of the distribution should be verified before using it.
~ For normally distributed scores
o Transform the X scores (values) into z-scores
o Look up the proportions corresponding to the z-score values
Chapter 7: Probability and Samples: The Distribution of
Sample Means
Shape, Central Tendency, and Variability for the Distribution of Sample Means:
• The central limit theorem is used to specify the shape, central tendency, and variability of the
distribution of sample means.
• The standard error of M measures how well an individual sample mean represents the entire
distribution.
• So, the central limit theorem is used to specific the shape, central tendency and variability of a
distribution of sample means.
• It provides a description of the distribution that would be obtained if you selected every possible
sample mean and constructed a distribution of the sample mean.
• Then the standard error of the mean measure how well an individual sample mean represents the
entire distribution.
Central Limit Theorem
• Applies to any population with mean μ and
standard deviation σ
• Distribution of sample means approaches a normal distribution as n approaches infinity
• Distribution of sample means for samples
of size n will have a mean of μM
s
• Distribution of sample means for samples of size n will have a standard deviation = n
• For any population with mean u and standard deviation o, the distribution sample means for
sample size n will have a mean of u and standard deviation of o/ square root of n and will
approach a normal distribution as n approaches infinity.
• So essentially he value of this theorem comes from the fact that it describes the distribution of
any sample means for any population, no matter what the shape, mean or standard deviation is.
• It also states that the distribution of the sample means approaches a normal distribution very
rapidly.
• So this theory follows three steps: shape, central tendency and variability…
Shape of the Distribution of Sample Means
• The distribution of sample means is almost perfectly normal in either of two conditions
• The population from which the samples are selected is a normal distribution
or
• The number of scores (n) in each sample is relatively large—at least 30
• Speaking to the shape of the distribution sample means, it tends to be normal if there are two
specific conditions:
1. the population from which the samples are selected is a normal distribution.
2. the number of scores (n) in each sample is relatively large, around 30 or more.
• So as n gets larger, the distribution of sample means will closely approximate normal distribution.
• So essentially as n gets greater than 30, the distribution is normal regardless of the shape of the
original population.
Expected Value of M
• The mean of the distribution of sample means is μM and has a value equal to the mean of the
population of scores, μ.
• The mean of the distribution of sample means is called the expected value of M.
• M is an unbiased statistic because μM, the expected value of the distribution of sample means, is
the value of the population mean, μ.
• So, as we said in the previous slide, the sample means is centred around the mean of the
population from which samples were obtained.
• We can then go one step further and say that in actual fact, the average value of all sample means
is exactly equal to the value of the population mean.
• This mean value is called the expected value of M.
• So in laymen’s terms, a sample mean is expected be near its population mean.
• Please also note the symbols here as well.
• So M is an unbiased statistic because on average the sample statistic produces a value that is
equal to the corresponding population.
Standard Error of M
• Variability of a distribution of scores is measured by the standard deviation.
• Variability of a distribution of sample means is measured by the standard deviation of the sample
means and is called the standard error of M and written as σM.
• In journal articles or other textbooks, the standard error of M might be identified as “standard
error,” “SE,” or “SEM”.
• Now we need to go one step further and speak about the third characteristic known as Variability.
• So the value we will be working with is the standard deviation for the distribution of sample
means. This is identified with the symbol here and is called standard error of M.
• Note that in other articles or textbooks , the standard error of M is symbolised or identified using
Standard error, SE or SEM.
Definition of the Standard Error of M
• The standard error of M is the standard deviation of the distribution of sample means. When it is
large, the means are widely scattered.
• The standard error of M provides a measure of how much distance is expected on average
between M and μ.
• The standard error of M serves the same two purposes for the distribution of sample means as we
spoke about in the previous podcast.
• Firstly, the SE describes the distribution of sample means.
• It provides a measure of how much difference is expected from one sample to another.
• So we note that when the SE is small, all the sample means are close together.
• When the SE is large, the sample means are widely scattered which means there are big
differences from one sample to another.
• Secondly, SE measures how well an individual sample mean represents the entire distribution.
• It provides a measure of how much distance is reasonable to expect between an sample mean
and the overall mean of the distribution.
• We should also remember that although a sample mean is representative of the population mean,
there typically is some error between the sample and the population.
• This is where the SE measures exactly how much error is expected on average between the
sample mean and the population mean.
Standard Error Magnitude
• Law of large numbers: the larger the sample size, the more probable it is that the sample mean
will be close to the population mean.
• Population variance: The smaller the variance in the population, the more probable it is that the
sample mean will be close to the population mean.
• So here we note the law of larger numbers states that the larger the sample size the more
probable it is that the sample mean will be close to the population mean.
• In terms of population variance, the smaller the variance in the population, the more probable it is
that the sample mean will be close to the population mean.
Figure 7.3 Standard Error and Sample Size Relationship
• The z-score tells us exactly where the sample mean is located in relation to all other possible
sample means that could have been obtained.
• So a z-score identifies the location with a signed number.
• The sign tells us whether the location is above or below the mean, we should know this by now.
• So if its positive then it’s above the mean, and if its negative it is below the mean.
• The number then also tells us the distance between the location and the mean in terms of the
number of standard deviations.
• Look here at the formula for calculating the z score for sample means: so z equals the M minus
the SD divided by the SD Error.
Calculating a z-Score for a sample mean
• The distribution of sample means for n = 16.
• Samples were selected from a normal population with μ = 500 and σ = 100.
Chapter 8: Introduction to Hypothesis Testing
The Logic of Hypothesis Testing
• Hypothesis testing is one of the most used inferential procedures.
• Definition: a statistical method that uses sample data to evaluate the validity of a hypothesis
about a population.
• State hypothesis about a population
• Predict the expected characteristics of the sample based on that hypothesis
• Obtain a random sample from the population
• Compare the obtained sample data with the prediction made from the hypothesis
• If consistent, hypothesis is reasonable
• If discrepant, hypothesis is rejected
Four Steps in Hypothesis Testing
Step 1: State the hypotheses
Step 2: Set the criteria for a decision
Step 3: Collect data & compute sample statistics
Step 4: Make a decision
Step 1: State Hypotheses
• Null hypothesis (H0) states that the treatment has no effect: there is no change, difference, or
relationship.
• The independent variable has no effect on the dependent variable.
• Alternative hypothesis (H1) states that there is a change, difference, or relationship in the general
population.
Step 2: Set the Decision Criterion
• Distribution of sample outcomes is divided
– Those likely if H0 is true
– Those “very unlikely” if H0 is true
• Alpha level, or significance level, is a probability value used to define “very unlikely” outcomes.
• Critical region(s) consist of the extreme sample outcomes that are “very unlikely.”
• Boundaries of critical region(s) are determined by the probability set by the alpha level.
Figure 8.3 Note “Unlikely” Parts of Distribution of Sample Means
o The set of potential samples is divided into those that are likely to be obtained and those that are
very unlikely to be obtained if the null hypothesis is true.
Figure 8.4 Critical region(s) for α = .05
s s2
sM = sM =
n n
Reviewing the z-Score Statistic
• Use z-score statistic to quantify inferences about the population.
• Use unit normal table to find the critical region if z-scores form a normal distribution
The Problem with z-Scores
• The z-score requires more information than researchers typically have available
• Requires knowledge of the population standard deviation σ
• Researchers usually have only the sample data available
Estimated standard error
• Use s2 to estimate σ2.
• Estimated standard error:
s s2
estimated standard error = sM =
or
n n
• Estimated standard error is used as estimate of the real standard error when the value of σM is
unknown.
Introducing the t Statistic
• t statistic is an alternative to z.
• t might be considered an “approximate” z.
• Estimated standard error (sM) is used as in place of the real standard error when the value of σM is
unknown.
The t-Statistic
• The t statistic uses the estimated standard error in place of σM.
M -µ
t=
sM
• The t statistic is used to test hypotheses about an unknown population mean μ when the value of
σ is also unknown.
Degrees of Freedom
• Computation of sample variance requires computation of the sample mean first
• Only n-1 scores in a sample are independent
• Researchers call n-1 the degrees of freedom
• Degrees of freedom
• Noted as df
• df = n − 1
Figure 9.1: Distributions of the t Statistic
• Distributions of the t statistic for different values of degrees of freedom are compared to a normal
z-score distribution.
• Like the normal distribution, t distributions are bell-shaped and symmetrical and have a mean of
zero.
• However, t distributions are more variable, indicated by the flatter and more spread-out shape.
• The larger the value of df is, the more closely the t distribution approximates a normal
distribution.
The t Distribution
• Family of distributions, one for each value of degrees of freedom
• Approximates the shape of the normal distribution
• Flatter than the normal distribution
• More spread out than the normal distribution
• More variability (“fatter tails”) in t distribution
• Use Table of Values of t in place of the Unit Normal Table for hypothesis tests
The t distribution table
Figure 9.2
• The basic research situation for using the t statistic hypothesis test. It is assumed that the
parameter μ is known for the population before treatment.
• The purpose of the research study is to determine whether the treatment has an effect.
• Note that the population after treatment has unknown values for the mean and the variance.
• We will use a sample to test a hypothesis about the population mean.
Hypothesis Testing: Four Steps
• State the null and alternative hypotheses and
select an alpha level
• Locate the critical region using the t distribution table and df
• Calculate the t test statistic
• Make a decision regarding H0 (null hypothesis)
Figure 9.4: Critical Region in the t Distribution for α = .05 and df = 8
s2
s22
= +
1
s( M - M 2)
1
n1 n2
Box 10.1: Variability of Difference Scores
• Why add sample measurement errors (squared deviations from mean) but subtract sample means
to calculate a difference score?
• Some instructors may want to advise students that the material in this box is designed to be
background support for understanding variability and that some students find it very insightful.
Figure 10.2 Two Population Distributions
• Degrees of freedom (df) for t statistic is df for first sample + df for second sample
df = df1 + df 2 = (n1 - 1) + (n2 - 1)
Note: this term is the same as the denominator of the pooled variance
Example
Chapter 11: The t Test for Two Related Samples
Introduction to Repeated-Measures Designs
• Repeated-measures design
• Also known as within-subjects design
• Two separate scores are obtained for each individual in the sample
• Same subjects are used in both treatment conditions
• No risk of the participants in each treatment group differing significantly from each other
Example
MD =
åD
• Mean Difference: n
= 9.6
M D - µD
t=
sM D
t = (5.5-9.6) / 2
t = -2.05
• For example, a researcher could check high school records (with permission) to obtain a measure of
each student’s academic performance, and then survey each family to obtain a measure of income.
• The resulting data could be used to determine whether there is a relationship between high school
grades and family income.
• Notice that the researcher is not manipulating any student’s grade or any family’s income, but is
simply observing what occurs naturally.
• You also should notice that a correlation requires two scores for each individual (one score from each
of the two variables).
• These scores normally are identified as X and Y.
• The pairs of scores can be listed in a table, or they can be presented graphically in a scatter plot
(Figure 14.1).
• In the scatter plot, the values for the X variable are on the horizontal axis and the Y values are listed
on the vertical axis.
• Each individual is represented by a single point in the graph so that the horizontal position
corresponds to the individual’s X value and the vertical position corresponds to the Y value.
• The value of a scatter plot is that it allows you to see any patterns or trends that exist in the data.
• The data points in Figure 14.1, for example, show a clear relationship between family income and
student grades; as income increases, grades also increase
Figure 14.2 Positive and Negative Relationships
• A recent report examined the relationship between income and weight for both men and women
(Judge & Cable, 2011).
• The results of the study are shown in Figure 14.2 and provide examples of positive and negative
relationships.
• For the men [Figure 14.2(a)], there is a positive relationship between weight and income, with income
increasing as weight increases from person to person.
• For the women [Figure 14.2(b)], the relationship is negative, with income decreasing as weight
increases.
• Although there is no definite explanation for the two relationships, the authors suggest that being
thin is seen as a sign of success and self-discipline for women.
• For men, however, a degree of extra weight is often viewed as a sign of success.
• FIGURE 14.2 Examples of positive and negative relationships.
(a) Income is positively related to weight for men.
(b) Income is negatively related to weight for women.
Figure 14.3 Different Linear Relationship Values
• FIGURE 14.3 Examples of difference values for linear correlations: (a) a perfect negative correlation, -
1.00; (b) no linear trend, 0.00; (c) a strong positive relationship, approximately +0.90; (d) a relatively
weak negative correlation, approximately -0.40.
The Pearson Correlation
• Measures the degree and the direction of the linear relationship between two variables
• Perfect linear relationship
• Every change in X has a corresponding change in Y
• Correlation will be −1.00 or +1.00
covariability of X and Y
r=
variability of X and Y separately
• By far the most common correlation is the Pearson correlation (or the Pearson product– moment
correlation), which measures the degree of linear relationship; that is, how well the data points fit
a straight line.
• The Pearson correlation for a sample is identified by the letter r.
• The corresponding correlation for the entire population is identified by the Greek letter rho (fancy
p), which is the Greek equivalent of the letter r.
• Conceptually, this correlation is computed by:
• When there is a perfect linear relationship, every change in the X variable is accompanied
by a corresponding change in the Y variable.
• In Figure 14.3(a), for example, every time the value of X increases, there is a perfectly
predictable decrease in the value of Y.
• The result is a perfect linear relationship, with X and Y always varying together.
• In this case, the coverability (X and Y together) is identical to the variability of X and Y
separately, and the formula produces a correlation with a magnitude of 1.00 or -1.00.
Sum of Products of Deviations (SP)
• Like SS (sum of squared deviations)
• Measures the amount of coverability between two variables
• SP definitional formula:
SP = å(X - M X )(Y - M Y )
• The calculation of the Pearson correlation requires one new concept: the sum of products of
deviations, or SP.
• This new value is similar to SS (the sum of squared deviations), which is used to measure
variability for a single variable.
• Now, we use SP to measure the amount of coverability between two variables.
• The value for SP can be calculated with either a definitional formula or a computational formula.
The definitional formula for the sum of products is___ where MX is the mean for the X scores and
MY is the mean for the Y scores.
• The definitional formula instructs you to perform the following sequence of operations: 1. Find the
X deviation and the Y deviation for each individual. 2. Find the product of the deviations for each
individual. 3. Add the products.
• Notice that this process literally defines the value being calculated; that is, the formula actually
computes the sum of the products of the deviations.
SP – Computational Formula
• Definitional formula emphasizes SP as the sum of two difference scores
• Computational formula results in easier calculations
• SP computational formula:
SP = å XY -
å X åY
n
• The computational formula for the sum of products of deviations is:
• Because the computational formula uses the original scores (X and Y values), it usually results in
easier calculations than those required with the definitional formula, especially if MX or MY is not
a whole number.
• However, both formulas will always produce the same value for SP.
• You may have noted that the formulas for SP are like the formulas you have learned for SS (sum
of squares).
• Specifically, the two sets of formulas have exactly the same structure, but the SS formulas use
squared values (X times X) and the SP formulas use products (X times Y).
Example
• The same set of n = 4 pairs of scores are used to calculate SP, first using the definitional formula,
and then using the computational formula.
• For the definitional formula, you need deviation scores for each of the X values and each of the Y
values. Note that the mean for the Xs is MX = 2.5 and the mean for the Ys is MY = 5. The
deviations and the products of deviations are shown in the following table:
• For the computational formula, you need the X value, the Y value, and the XY product for each
individual. Then you find the sum of the Xs, the sum of the Ys, and the sum of the XY products.
These values are as follows:
• Before starting any calculations, it is useful to put the data in a scatter plot and make a preliminary
estimate of the correlation.
• These data have been graphed in Figure 14.4 in your textbook.
• Looking at the scatter plot, it appears that there is a very good (but not perfect) positive
correlation.
• You should expect an approximate value of r = +.8 or +.9.
• To find the Pearson correlation, we need SP, SS for X, and SS for Y.
• The calculations for each of these values, using the definitional formulas, are presented in the
table below. (Note that the mean for the X values is MX = 6 and the mean for the Y scores is MY =
4.)
• Note that the correlation accurately describes the pattern shown in Figure 14.4.
• The positive sign is consistent with the fact that the Y values increase as X increases from left to right
in the graph.
• Also, if a line were drawn through the data points, from the bottom left corner where the two axes
meet to the top right data point (X = 10 and Y = 6), the data points would be very close to the line,
indicating a very good correlation (near 1.00).
Pearson Correlation and z-Scores
• Pearson correlation formula can be expressed as a relationship of z-scores.
Sample : r =
åz z
X Y
n -1
Population : r =
åz z
X Y
N
• The Pearson correlation measures the relationship between an individual’s location in the X
distribution and his or her location in the Y distribution.
• For example, a positive correlation means that individuals who score high on X also tend to
score high on Y.
• Similarly, a negative correlation indicates that individuals with high X scores tend to have
low Y scores.
• Recall from Chapter 5 that z-scores identify the exact location of each individual score within a
distribution.
• Each X value can be transformed into a z-score, zX, using the mean and standard deviation for the
set of Xs.
• Similarly, each Y score can be transformed into zY.
• If the X and Y values are viewed as a sample, the transformation is completed using the sample
formula for z (Equation 5.3, page 124).
• If the X and Y values form a complete population, the z-scores are computed using Equation 5.1
(page 122).
• After the transformation, the formula for the Pearson correlation can be expressed entirely in
terms of z-scores.
Correlations and Outliers
• An outlier is an extremely deviant individual in the sample
• Characterized by a much larger (or smaller) score than all the others in the sample
• In a scatter plot, the point is clearly different from all the other points
• Outliers produce a disproportionately large impact on the correlation coefficient
• An outlier is an individual with X and/or Y values that are substantially different (larger or smaller)
from the values obtained for the other individuals in the data set.
• The data point of a single outlier can have a dramatic influence on the value obtained for the
correlation.
• The problem of outliers is a good reason for looking at a scatter plot instead of simply basing your
interpretation on the numerical value of the correlation.
• If you only “go by the numbers,” you might overlook the fact that one extreme data point inflated
the size of the correlation.
Figure 14.7 Outlier Influences Size of Correlation
• FIGURE 14.7 A demonstration of how one extreme data point (an outlier) can influence the value of a
correlation.
Correlation and Causation
• Correlation describes a relationship but does not demonstrate causation.
• Establishing causation requires an experiment in which one variable is manipulated and others
carefully controlled.
• Example 14.4 (and Figure 14.5) demonstrates the fallacy of attributing causation after observing a
correlation.
Figure 14.5 Correlation: Churches and Serious Crimes
• Hypothetical data showing the logical relationship between the number of churches and the number
of serious crimes for a sample of U.S. cities.
14.4 Hypothesis Tests with the Pearson Correlation
• Pearson correlation is usually computed for sample data but is also used to test hypotheses about
the relationship in the population.
• Population correlation is shown by Greek letter rho (ρ).
• Non-directional: H0: ρ = 0 and H1: ρ ≠ 0
Directional: H0: ρ ≤ 0 and H1: ρ > 0 or Directional: H0: ρ ≥ 0 and H1: ρ < 0
• The Pearson correlation is generally computed for sample data. As with most sample statis- tics,
however, a sample correlation is often used to answer questions about the corresponding
population correlation.
• For example, a psychologist would like to know whether there is a relationship between IQ and
creativity.
• This is a general question concerning a population.
• To answer the question, a sample would be selected, and the sample data would be used to
compute the correlation value.
• You should recognize this process as an example of inferential statistics: using samples to draw
inferences about populations.
• In the past, we have been concerned primarily with using sample means as the basis for answering
questions about population means.
• In this section, we examine the procedures for using a sample correlation as the basis for testing
hypotheses about the corresponding population correlation.
• The basic question for this hypothesis test is whether a correlation exists in the population.
• The null hypothesis is “No. There is no correlation in the population” or “The population
correlation is zero.”
• The alternative hypothesis is “Yes. There is a real, nonzero correlation in the population.”
Because the population correlation is traditionally represented by P (the Greek letter rho),
these hypotheses would be stated in symbols as
• When there is a specific prediction about the direction of the correlation, it is possible to do a
directional, or one-tailed test.
• For example, if a researcher is predicting a positive relationship, the hypotheses would be
• The correlation from the sample data is used to evaluate the hypotheses.
• For the regular, nondirectional test, a sample correlation near zero provides support for
H0 and a sample value far from zero tends to refute H0.
• For a directional test, a positive value for the sample correlation would tend to refute a
null hypothesis stating that the population correlation is not positive.
Figure 14.9 Correlation in Sample vs. Population
• Although sample correlations are used to test hypotheses about population correlations, you
should keep in mind that samples are not expected to be identical to the populations from which
they come; there will be some discrepancy (sampling error) between a sample statistic and the
corresponding population parameter.
• Specifically, you should always expect some error between a sample correlation and the
population correlation it represents.
• One implication of this fact is that even when there is no correlation in the population (r = 0), you
are still likely to obtain a nonzero value for the sample correlation.
• This is particularly true for small samples.
• Figure 14.9 illustrates how a small sample from a population with a near-zero correlation could
result in a correlation that deviates from zero.
• The coloured dots in the figure represent the entire population and the three circled dots
represent a random sample.
• Note that the three sample points show a relatively good, positive correlation even through there
is no linear trend (r = 0) for the population.
• When you obtain a nonzero correlation for a sample, the purpose of the hypothesis test is to
decide between the following two interpretations:
1. There is no correlation in the population (r = 0) and the sample value is the result of sampling
error.
Remember, a sample is not expected to be identical to the population.
There always is some error between a sample statistic and the corresponding population
parameter.
This is the situation specified by H0.
2. The nonzero sample correlation accurately represents a real, nonzero correlation in the
population.
This is the alternative stated in H1.
• The correlation from the sample will help to determine which of these two interpretations is more
likely.
• A sample correlation near zero supports the conclusion that the population correlation is also zero.
• A sample correlation that is substantially different from zero supports the conclusion that there is a
real, nonzero correlation in the population.
Correlation Hypothesis Test
• Sample correlation r used to test population ρ
• Degrees of freedom (df) = n − 2
• Hypothesis test can be computed using
either t or F; only t shown in this chapter
• Use t table to find critical value with df = n − 2
r-r
t=
(1 - r 2 )
(n - 2)
• The hypothesis test evaluating the significance of a correlation can be conducted using either a t
statistic or an F-ratio.
• The F-ratio is discussed later (pages 460–462) and we focus on the t statistic here.
• The t statistic for a correlation has the same general structure as t statistics introduced in
Chapters 9, 10, and 11. (sample statistic – population parameter / standard error)
• In this case, the sample statistic is the sample correlation (r) and the corresponding parameter is
the population correlation (P).
• The null hypothesis specifies that the population correlation is r = 0.
• The final part of the equation is the standard error, which is determined by
• Thus, the complete t statistic is FORMULA
• Degrees of Freedom:
• The t statistic has degrees of freedom defined by df = n - 2. An intuitive explanation for
this value is that a sample with only n = 2 data points has no degrees of freedom.
• Specifically, if there are only two points, they will fit perfectly on a straight line, and the
sample produces a perfect correlation of r = +1.00 or r = -1.00.
• Because the first two points always produce a perfect correlation, the sample correlation is
free to vary only when the data set contains more than two points.
• Thus, df = n - 2.
Example