Download as pdf or txt
Download as pdf or txt
You are on page 1of 113

EDUC 707 PORTFOLIO

Jessica Buckle

WINTER 2020
CALIFORNIA STATE UNIVERSITY, SAN BERNARDINO
1

Contents
Assignment 1, Terminology and Statistical Tests.......................................................................................... 2
Assignment 2, Terminology and Central Tendency ...................................................................................... 8
Assignment 3, T-Tests ................................................................................................................................. 19
Assignment 4, Terminology, Research Questions, Hypothesis Testing and Assumptions of Parametric
Statistics ...................................................................................................................................................... 32
Assignment 5, Terminology, Significance, Type I and II Errors, Confidence Intervals, Effect Size, Power, Z-
Test.............................................................................................................................................................. 41
Assignment 6, Terminology, ANOVA, F-Statistic......................................................................................... 47
Assignment 7, Terminology, Correlations, Reliability, Validity, and Generalizability ................................. 62
Assignment 8, Terminology, Correlations, and Linear Regression ............................................................. 87
Assignment 9, Terminology, Nonparametric Statistics, and Chi-Square .................................................. 103
2

Jessica Buckle
EDUC 707 W20

EDUC 707

Assignment 1, Terminology and Statistical Tests


Define the following terms and answer the questions:
1. Descriptive Statistics – Descriptive statistics are used to organize and describe the
characteristics of a collection of data
2. What are some examples of descriptive data?
Examples of descriptive data might be gender, age, years of experience, grade
point average, occupation, ethnicity, major in school, etc.

3. Inferential Statistics – Inferential statistics are generally (but not always) the next step after
collecting and summarizing data. It is used to make inferences based on s smaller group of data,
usually called a sample (portion/subset) of a population.
4. What are some examples of inferential data?
Examples of inferential data such as making generalizations about all fifth
graders in the state of New York after using a sample of data collected from 150
students or making generalizations about women who shop at Target after using a
sample of data collected from 200 women who have shopped at Target.
5. Sample – A sample is a portion or subset of a population.
6. What is an example of a sample?
Examples of a sample would be a group of students at a school, a group of
workers at a business or in an occupational field, or a group of shoppers that
frequent a store or buy a certain product.

7. Population – A population is all possible subjects or cases of interest.


8. What is an example of a population?
Examples of populations would be teenagers, women, video game players,
teachers, first born children, etc.
9. What is central tendency? Central tendency is a central or typical value for a probability
distribution, generally being called averages- mean, median, and mode.
10. Mean – The mean is the most common type of average that is computed. It is the sum of all
the values in a group, divided by the number of values in that group. So if you had the test scores
of 30 students you would add them up to get the total, then divide the total by 30 to get the mean.
11. Median – The median is the midpoint in a set of scores. It is the point at which half of the
scores fall above and the other half fall below. If you have the following scores: 92, 87, 76, 65,
3

and 52 the median would be 76. If you have an even number of values then the median is the
mean of the two middle values. For example, with the values 51, 43, 34, 32, 27, and 12 the
median is 33.
12. Mode – The mode is the value that occurs most frequently. For example if there is a multiple
choice question and 57 people pick A, 20 people pick B, 12 people pick C, and 11 people pick D
the modal response is A since it was the answer that was selected the most.
13. What do parametric statistics assume? Parametric statistics assume the variance of each
group are similar and that the sample is large enough to represent the population- it takes a
sample size of about 30 (according to the book) to fulfill this assumption. We have a normal
distribution that results in a bell curve.
14. List 3 examples of parametric tests:
Examples of parametric tests are the Paired t-test, Unpaired t-test, and Pearson
correlation

15. What do nonparametric statistics NOT assume? Nonparametric statistics do not follow
require the same, possibly restrictive assumptions as the parametric test but it does NOT assume
there are no assumptions.
16. List 3 versions of nonparametic tests that align with the parametric tests listed
for number 14:
The nonparametric tests that align with the above parametric tests from #14 are
the Wilcoxon Rank sum Test, the Mann-Whitney U test, and the Spearman correlation
17. What are Nominal, Ordinal, Interval, and Ratio data? Please provide a definition and
example of each.
Nominal – The nominal level of measurement is defined by the characteristics of an
outcome that fit into one and only one class or category. They can be the least precise level of
measurement. Examples are gender (female and male), ethnicity (Caucasian or African
American), or political affiliation (Democrat, Republican, or Independent)
Ordinal – The ordinal level of measurement examines things that are ordered. Examples
would be rank of candidates for a job or a rank of students based on gpa.
Interval – The interval level of measurement is a test or assessment tool based on some
underlying continuum wherein we can talk about how much more something is or less than
something is (versus just knowing the order or things). The intervals or points along the scale are
equal to one another. An example would be a spelling test where you get 10 answers correct. 10
is twice as many as only getting 5 correct and it is 2 more than only getting 8 correct.
Ratio – The ratio level of measurement is characterized by the presence of an absolute
zero on the scale (meaning there could be an absence of the trait being measured. Examples
include age, height, weight, and years of education.
18. Z-test –
4

What is a Z-test used for? The Z-test is used to determine whether two population means are
different when the variances are known and the sample size is large (greater than 30).
What are the assumptions of a Z-test?
The assumptions of the Z-test are:
• Interval or ratio scale of measurement (approximately interval)
• Random sampling from a defined population
• Characteristic is normally distributed in the population

What is the Null Hypothesis (H0) for a Z-test?


H0: x = μ (the sample mean is the same as the population mean)

What is the Alternative Hypothesis (Ha) for a Z-test?


Ha: x ≠ μ (the sample mean is not the same as the population mean)

What is an example of a basic question that could be answered using a Z-test?


RQ: A principal at a certain school claims that the students in his school are above
average intelligence. A random sample of thirty students IQ scores have a mean score of 112. Is
there sufficient evidence to support the principal’s claim? The mean population IQ is 100 with a
standard deviation of 15.
H0: μ=100 (population mean equals 100)
Ha: μ > 100 (the claim is that students have above average IQ scores)

EXAMPLE: 19. T-test –


What is a T-test used for?
A t-test is a type of inferential statistic used to determine if there is a significant
difference between the means of two groups.
What are the three versions of a t-test?
Independent samples t-test compares the means for two different groups.
5

Paired (dependent) samples t-test compares two different means from the same group.
One sample t-test tests the mean from one group against a known mean.
What are the assumptions of a T-test?
Data is on a continuous or ordinal scale.
Random sample
Normal distribution
Large enough sample to achieve a bell-shaped curve (i.e. normal distribution)
Homogeneity of variance (i.e. equal variance, the standard deviation from the samples are
approximately equal.)
What is the Null Hypothesis (H0) for a T-test?
One-tail t-test:
H0: µ ≤ a known mean (i.e. sample mean is less than or equal to a known mean)
H0: µ ≥ a known mean (i.e. sample mean is greater than or equal to a known mean)
Two-tail t-test:
H0: µ1 = µ2 (i.e. sample 1 mean is equal to sample two mean)
What is the Alternative Hypothesis (Ha) for a T-test?
One-tail t-test:
Ha: µ > a known mean (i.e. sample mean is greater than a known mean)
Ha: µ < a known mean (i.e. sample mean is less than a known mean)
Two-tail t-test:
Ha: µ1 ≠ µ2 (i.e. sample 1 mean is not equal to sample two mean)
What is an example of a basic question that could be answered using a T-test?
RQ: Did students in Dr. Hughes’ statistics class perform better than the students in the
same course previously?
This is known as a one-sided, directional, or one-tailed t-test. The words “performed
better” indicate a direction.
H0: µ ≥ a known mean
Ha: µ < a known mean
RQ: Did students in Dr. Hughes’ statistics class perform the same as the students in same
course previously?
6

This is known as a two-sided, non-directional, or two-tailed t-test. This would only


indicate whether students performed the same or not the same, and NOT whether the
students performed better or worse.
H0: µ1 = µ2
Ha: µ1 ≠ µ2
20. ANOVA –
What is an ANOVA used for? ANOVA is used to compare multiple (three or more) samples
with a single test.
What are the two primary (basic) versions of an ANOVA?
1. One-way ANOVA: used to compare the difference between the three or more
samples/groups of a single independent variable
2. MANOVA: used to test the effect of one or more independent variable on two or more
dependent variables. In addition, MANOVA can also detect the difference in co-relation
between dependent variables given the groups of independent variables.

What is the Null Hypothesis (H0) for an ANOVA?


H0: µ1 = µ2 = µ3…= µk (all pairs of samples are the same- i.e. all sample means are equal)

What is the Alternative Hypothesis (Ha) for an ANOVA?


Ha: The means are not all equal (at least one pair of samples is significantly different)

What is an example of a basic question that could be answered using an ANOVA?


RQ: A group of psychiatric patients are trying three different therapies: counseling,
medication and biofeedback. You want to see if one therapy is better than the others.

H0: µ1 = µ2 = µ3 (the three therapies are equal in their effectiveness)

Ha: µ1 > µ2 (One of the therapies is significantly better than the others)

21. Chi-Square –
What is a Chi-Square Test used for? The Chi-Square test is used to compare categorical
variables.
What are the two primary types of a Chi-Square Test?
1. Goodness of fit test, which determines if a sample matches the population
2. A chi-square fit test for two independent variables in a contingency table to check if the
data fits
a. A Small chi-square value means the data fits
b. A high chi-square value means that data does not fit

What is the Null Hypothesis (H0) for a Chi-Square Test?


H0: Variable A and Variable B are independent
7

What is the Alternative Hypothesis (Ha) for a Chi-Square Test?


Ha: Variable A and Variable B are not independent

What is an example of a basic question that could be answered using a Chi-Square Test?
RQ: Does gender affect the preference of the type of vacation?
8

Jessica Buckle
Winter 2020

EDUC 707

Assignment 2, Terminology and Central Tendency


Define the following terms and answer the questions:
1. Compute the mean of 9, 8, 6, 5, 3, and 2 (show all work):
The mean is the average of the numbers so you add the numbers together and then divide
by how many numbers there are. In this instance it would be 9+8+6+5+3+2 = 33/6= 5.5 so the
mean of this data set would be 5.5.
2. Determine the median of 9, 8, 6, 5, 3, and 2 (explain of you determined the median):
The median is the number that is halfway into the set. To find the median, the data should
be arranged in order from least to greatest. If there is an even number of items in the data set,
then the median is found by taking the mean (average) of the two middlemost numbers. In this
instance, since there are an even number of items you’d take the two middle numbers (in this
case 6 and 5 and determine the mean- 6+5=11/2=5.5) so the median for this data set would be
5.5.
3. What are percentile points?
A percentile point is the percentage of cases equal to and below a particular score in a
distribution or set of scores. For example, if a score is at the 75th percentile, it means that the
score is at or above 75% of the other scores in the distribution.
9

4. Provide at least two examples of percentile points from the data in questions 1 and 2.
If we are given the data set of 9, 8, 6, 5, 3, and 2 we know that 5.5 (the median) is the 50th
percentile (also known as Q2) - below this point is where 50% of the distribution of cases falls.
From there, if we divide our data set into quarters it will give us further percentile points- a score
of 3 would mark the 25th percentile (or Q1) and a score of 8 would mark the 75th percentile (or
Q3).
5. Explain quartiles?
Quartiles are similar to percentile points in that they divide the data into four quarters and
give us percentile points.
10

6. What quartiles are the examples from question 4 in?


If we are given the data set of 9, 8, 6, 5, 3, and 2 we know that 5.5 (the median) is the 50th
percentile (also known as Q2) - below this point is where 50% of the distribution of cases falls.
From there, if we divide our data set into quarters it will give us further percentile points- a score
of 3 would mark the 25th percentile (or Q1) and a score of 8 would mark the 75th percentile (or
Q3).
7. What is the mode in this data 9, 9, 8, 6, 6, 5, 3, 3, 3, and 2:
The mode of a data set is the number that occurs most frequently in the set. To easily find
the mode, put the numbers in order from least to greatest and count how many times each
number occurs. The number that occurs the most is the mode. In the instance of the data set
provided above 3 would be the mode as it occurs more times than any other number.
8. What are the modes in this data 9, 9, 8, 6, 6, 6, 5, 3, 3, 3, and 2:
11

In this instance there are two modes because 6 and 3 both occur three times, which is the
most number of times. So in this data set the mode is 3 and 6.
9. What is skew?
A skew is the quality of a distribution that defines the disproportionate frequency of
certain scores. A longer right tail than left corresponds to a smaller number of occurrences at the
high end of the distribution (which would be a positive skew). A shorter right tail than left
corresponds to a larger number of occurrences at the high end of the distribution (which would
be a negative skew).
12

10. What does skew indicate?


Skew indicates where the outliers are and where a greater distribution of scores lies. If the
skew=0 the data is symmetrical. If the skew > 0 the distribution is positively skewed and more
data is to the left. If the skew < 0 the distribution is negatively skewed and more data is to the
right.
11. Draw graphs of positive and negative skew:

12. What is variability?


Variability is how much scores differ from one another or the amount of spread or
dispersion in a set of scores. There are multiple ways to measure variability such as range,
quartiles, variance, and standard deviation.
13. What is the range of this data 9, 8, 6, 5, 3, and 2 (show all work):
13

The range is the most general measure of variability. It gives you an idea of how far apart
scores are from one another. It is computed by subtracting the lowest score in a distribution from
the highest score in the distribution. The formula is r= h-l so with the data set above we would
take 9 (which is the highest score) and subtract 2 (which is the lowest score) to get a range of 7
(9-2=7).
14. What is the standard deviation?
The Standard Deviation is the average distance for each score from the mean or how
concentrated the data is around the mean (a smaller SD indicated more concentration around the
mean). For example, if the mean is 90 points and the standard deviation of students test scores
out of 100 points is 1 point, 68% of students test scores are within 1 point of 90 points, and 95%
of students test scores are within 2 points of 90 points.
14

15. Calculate the standard deviation of this data 9, 8, 6, 5, 3, and 2 (show all work):

Step 1: Work out the mean- 9+8+6+5+3+2=33, 33/6=5.5 so the mean is 5.5
15

Step 2: For each number subtract the mean and square the result- (9-5.5=3.5^2=12.25), (8-
5.52.5^2=6.25), (6-5.5=.5^2=.25), (5-5.5=-.5^2=.25), (3-5.5=-2.5^2=6.25), (2-5.5=-3.5^2=12.25)
Step 3: work out the mean of those squared differences- (12.25+6.25+.25+.25+6.25+12.25-37.5)
divided by n-1(6-1=5), 37.5/5=7.5
Step 4: Take the square root- square root of 7.5 = 2.73861 so that is the Standard Deviation
16. What is the mean deviation?
The mean deviation is how far, on average, all values are from the middle.

17. What is the mean deviation of this data -9, 8, 6, -5, 3, and -2 (show all work and
explain):

Step 1: Find the mean- (-9 + 8+ 6+-5+ 3+-2=1, 1/6=.1666)


Step 2: Find the distance of each value from the mean- data (-9-.1666=9.1666), (8-.1666=6.334),
(6-.1666=5.8334), (-5-.1666=5.1666), (3-.1666=2.8334), and (-2-.1666=2.1666)
Step 3: Find the mean of those distances-(9.1666+6.334+5.8334+5.1666+2.8334+2.1666=
31.5006/6= 5.2501
18. What does it mean that the standard deviation is an unbiased estimate? And what does
that have to do with n-1?
The standard deviation is unbiased because we are estimating the population’s standard
deviation. We subtract 1 (n-1 in the equation) in order to make the standard deviation larger,
which makes the standard deviation more conservative.
16

19. What are the similarities and differences between variance and standard deviation?
Standard Deviation and Variance are similar in that they are both measures of variability,
dispersion, and spread. The formulas used to compute them are also similar. They are different in
several ways- the most important of which is that standard deviation is stated in the original units
from which it was derived (because we take the square root) whereas in variance the units stated
are squared.
17

20. What is the variance of this data 9, 8, 6, 5, 3, and 2 (show all work):
18

Step 1: Work out the mean- (9+ 8+ 6+ 5+ 3+ 2=33, 33/6=5.5)


Step 2: For each number subtract the mean and square the result- (9-5.5=3.5^2=12.25), (8-
5.5=2.5^2=6.25), (6-5.5=.5^2=.25), (5-5.5=-.5^2=.25), (3-5.5=-2.5^2= 6.25), (2-5.5=-
3.5^2=12.25)
Step 3: Work of the average of those squared differences- 12.25+6.25+.25+.25+6.25+12.25=37.5
-37.5/6=6.25
21. What are the 4 ways that a distribution can vary? Please define each way.
22. Draw pictures of each of the 4 ways that distributions can vary.
23. What is Kurtosis?
Kurtosis is the quality of a distribution that defines how flat or peaked it is.

24. What is platykurtic?


Platykurtic is a form of kurtosis where kurtosis is less than zero so the distribution is
more flattened. In the picture above it would be the red lined example.
25. What is leptokurtic?
Leptokurtic is a form of kurtosis where kurtosis is greater than zero so the distribution is
taller and skinner. In the picture above, it would be the blue lined example.
19

Jessica Buckle
EDUC 707
Meeting 3 HW

Assignment 3, T-Tests
Research Question 1

Section 1: Research Questions: In this section you will write 5 research questions based on the data
provided in Blackboard
Do male professors make the same salary as female professors at the same university?

Section 2: Null and Alternative Hypotheses: In this section you will write Null and Alternative
Hypotheses for T-tests based on your research questions and the data provided in Blackboard
Null Hypothesis: Male professors’ salaries ARE equal to female professors’ salaries.

Alternative Hypothesis: Male professors’ salaries are NOT equal to female professors’ salaries

Section 3: Check your assumptions and present evidence to suggest your check of the assumptions
(Note: Assumption videos in presentation)
Test for Normality: For this research question, the assumption is that the null hypothesis
distributions are normal. According to the Shapiro-Wik test of normality, which provides us with p-
values of .805 and .847 (Table 1), we can assume the distributions are normal because both values are
above .05.

Table 1
20

Homogeneity of Variance: For this research question, the assumption is that variances are

equal. According to the Shapiro-Wik test for homogeneity of variance, we are provided with a p-value

of .831 (Table 2) between groups, which is greater than .05, so we accept the null hypothesis that the

variances are equal.

Table 2

Because both the test for normality and the test for homogeneity of variance confirm our assumptions,
we can run the t-test.

Section 4: Run the t-tests (even if the assumptions are not upheld) to answer your hypotheses Because
both the test for normality and the test for homogeneity of variance confirm our assumptions,
we can run the t-test.

Section 5: Present the results of the t-tests using text and tables
21

T-Test: After running the t-test, we are provided with the p-value of .213 (Table 3).
Because this is above .05 we can accept the null hypothesis that the salaries of male and female
professors are equal.

Table 3

Section 6: Interpret the results to answer your research questions


Our research question was do male professors make the same salary as female professors at
the same university? Before we ran our t-test, we checked our assumptions by running a test of
normality and a test of homogeneity of variance. Based on the p-values in each test being above the
demarcation point of .05 (.805 and .847 for the test of normality and .831 for the homogeneity of
variance), our assumptions were proven to be true, which allowed us to run the t-test. After running
the t-test, we were given a p-value of .213, which is above the demarcation point of .05. Therefore, we
can accept the null hypothesis that Male professors’ salaries ARE equal to female professors’ salaries.
Thus, the answer to our research question is yes, male professors make the same salary as female
professors at the same university.

Research Question 2

Section 1: Research Questions: In this section you will write 5 research questions based on the data
provided in Blackboard
Is the percentage of Democratic and Republican voters equal in the recent national election?

Section 2: Null and Alternative Hypotheses: In this section you will write Null and Alternative
Hypotheses for T-tests based on your research questions and the data provided in Blackboard
22

Null Hypothesis: The percentage of Democratic voters IS equal to the percentage of

Republican voters in the recent national election.

Alternative Hypothesis: The percentage of Democratic voters is NOT equal to the percentage of

Republican voters in the recent national election.

Section 3: Check your assumptions and present evidence to suggest your check of the assumptions
(Note: Assumption videos in presentation)
Test for Normality: For this research question, the assumption is that the null hypothesis
distributions are normal. According to the Shapiro-Wik test of normality, which provides us with p-
values of .134 and .466 (Table 4), we can assume the distributions are normal because both values are
above .05.

Table 4

Homogeneity of Variance: For this research question, the assumption is that variances are
equal. According to the Shapiro-Wik test for homogeneity of variance, we are provided with a p-value
of .931 between groups (Table 5), which is greater than .05, so we accept the null hypothesis that the
variances are equal.

Table 5
23

Because both the test for normality and the test for homogeneity of variance confirm our assumptions,
we can run the t-test.

Section 4: Run the t-tests (even if the assumptions are not upheld) to answer your hypotheses
Because both the test for normality and the test for homogeneity of variance confirm our
assumptions, we can run the t-test.

Section 5: Present the results of the t-tests using text and tables
After running the t-test, we are provided with the p-value of .947 (Table 6). Because this is above .05
we can accept the null hypothesis that the percentage of democratic voters is equal to the percentage
of Republican voters.
Table 6

Section 6: Interpret the results to answer your research questions

Our research question was do male professors make the same salary as female professors at
the same university? Before we ran our t-test, we checked our assumptions by running a test of
normality and a test of homogeneity of variance. Based on the p-values in each test being above the
demarcation point of .05 (.134 and .466 for the test of normality and .931 for the homogeneity of
variance), our assumptions were proven to be true, which allowed us to run the t-test. After running
the t-test, we were given a p-value of .947, which is above the demarcation point of .05. Therefore, we
can accept the null hypothesis that the percentage of Democratic voters is equal to the percentage of
Republican voters in the recent national election. Thus, the answer to our research question is yes, the
24

percentage of Democratic voters is equal to the percentage of Republican voters in the recent national
election.

Research Question 3

Section 1: Research Questions: In this section you will write 5 research questions based on the data
provided in Blackboard
Is the height of basketball players the same as the height of football players?

Section 2: Null and Alternative Hypotheses: In this section you will write Null and Alternative
Hypotheses for T-tests based on your research questions and the data provided in Blackboard Null
Hypothesis: The height of basketball players IS the same as the height of football
players.

Alternative Hypothesis: The height of basketball players is NOT the same as the height of football

players.

Section 3: Check your assumptions and present evidence to suggest your check of the assumptions
(Note: Assumption videos in presentation)

Test for Normality: For this research question, the assumption is that the null hypothesis
distributions are normal. According to the Shapiro-Wik test of normality, which provides us with p-
values of .043 and .320 (Table 7), we can assume the distributions are NOT normal because one of the
values is below .05.

Table 7
25

Homogeneity of Variance: For this research question, the assumption is that variances are
equal. According to the Shapiro-Wik test for homogeneity of variance, we are provided with a p-value
of .0004078 between groups (Table 8), which is less than .05, so we must reject the null hypothesis that
the variances are equal because there is a significant statistical difference.

This assumption has been violated.

Table 8

The tests for normality and homogeneity of variance demonstrate that are assumptions have been
violated as there is not a normal distribution, nor are the variances equal. However, despite this we will
still run a t-test.

Section 4: Run the t-tests (even if the assumptions are not upheld) to answer your hypotheses
The tests for normality and homogeneity of variance demonstrate that are assumptions have
been violated as there is not a normal distribution, nor are the variances equal. However, despite this
we will still run a t-test.

Section 5: Present the results of the t-tests using text and tables
If we look at the p-value for the equal variances not assumed (due to the fact that the test for
homogeneity of variance proved the variances were not equal), we see the significance of .000 (Table
9). Because this value is below the demarcation point of .05 we must reject the null hypothesis that the
height of basketball players and football players is the same because there is a statistically significant
difference.

Table 9
26

Section 6: Interpret the results to answer your research questions

Our research question was is the height of basketball players the same as the height of football
players? Before we ran our t-test, we checked our assumptions by running a test of normality and a
test of homogeneity of variance. Based on the p-values in each test being below the demarcation point
of .05 (.043 and .320 for the test of normality and .0004078 for the homogeneity of variance), our
assumptions were rejected because there was not a normal distribution nor were the variances equal,
however we ran the t-test anyways. After running the t-test, we were given a p-value of .000, which is
below the demarcation point of .05. Therefore, we must reject the null hypothesis that the height of
basketball players is equal to the height of football players and accept the alternative hypothesis that
the height of basketball players is not equal to the height of football players . Thus, the answer to our
research question is no, the height of basketball players is not the same as football players.

Research Question 4

Section 1: Research Questions: In this section you will write 5 research questions based on the data
provided in Blackboard
Is the birth rate in California equal to the birth rate in Maine?

Section 2: Null and Alternative Hypotheses: In this section you will write Null and Alternative
Hypotheses for T-tests based on your research questions and the data provided in Blackboard
Null Hypothesis: The birth rate in California IS equal to the birth rate in Maine.

Alternative Hypothesis: The birth rate in California is NOT equal to the birth rate in Maine.

Section 3: Check your assumptions and present evidence to suggest your check of the assumptions
(Note: Assumption videos in presentation)
27

Test for Normality: For this research question, the assumption is that the null hypothesis
distributions are normal. According to the Shapiro-Wik test of normality, which provides us with p-
values of .995 and .504 (Table 10), we can assume the distributions are normal because both values are
above .05.

Table 10

Homogeneity of Variance: For this research question, the assumption is that variances are
equal. According to the Shapiro-Wik test for homogeneity of variance, we are provided with a p-value
of .000252797 (Table 11) between groups, which is less than .05, so we must reject the null hypothesis
that the variances are equal because there is a significant statistical difference.

This assumption has been violated.

Table 11
28

The tests for normality shows that our assumption of normal distribution is correct but the test for
homogeneity of variance demonstrated that the assumption of equal variances has been violated.
However, despite this we will still run a t-test.

Section 4: Run the t-tests (even if the assumptions are not upheld) to answer your hypotheses
The tests for normality shows that our assumption of normal distribution is correct but the test
for homogeneity of variance demonstrated that the assumption of equal variances has been violated.
However, despite this we will still run a t-test.

Section 5: Present the results of the t-tests using text and tables
If we look at the p-value for the equal variances not assumed (due to the fact that the test for
homogeneity of variance proved the variances were not equal), we see the significance of .000 (Table
12). Because this value is below the demarcation point of .05 we must reject the null hypothesis that
the birth rate in California is equal to the birth rate in Maine.

Table 12

Section 6: Interpret the results to answer your research questions

Our research question was is the birth rate in California equal to the birth rate in Maine? Before we ran
our t-test, we checked our assumptions by running a test of normality and a test of homogeneity of
variance. Based on the p-values of .995 and .504 for the test of normality we concluded that the
distributions were equal, therefore our assumption was correct. However, when we ran the
29

homogeneity of variance test we got the p-value of .000252797,which meant we had to reject our
assumption of the variances being equal as there is a significant statistical difference, however we ran
the t-test anyways. After running the t-test, we were given a p-value of .000, which is below the
demarcation point of .05. Therefore, we must reject the null hypothesis that the birth rate in California
is equal to the birth rate in Maine and accept the alternative hypothesis that the birth rate in California
is not equal to the birth rate in Maine . Thus, the answer to our research question is no, the birth rate
in California is not equal to the birth rate in Maine.

Research Question 5

Section 1: Research Questions: In this section you will write 5 research questions based on the data
provided in Blackboard
Are the scores of students in Dr. Hughes’ class equal to the scores of students in Dr.
Made-up’s class?
Section 2: Null and Alternative Hypotheses: In this section you will write Null and Alternative
Hypotheses for T-tests based on your research questions and the data provided in Blackboard Null
Hypothesis: The scores of students in Dr. Hughes’ class ARE equal to the scores of
students in Dr. Made-Up’s class.

Alternative Hypothesis: The scores of students in Dr. Hughes’ class are NOT equal to the scores of

students in Dr. Made-Up’s class.

Section 3: Check your assumptions and present evidence to suggest your check of the assumptions
(Note: Assumption videos in presentation)
Test for Normality: For this research question, the assumption is that the null hypothesis
distributions are normal. According to the Shapiro-Wik test of normality, which provides us with p-
values of .423 and .037 (Table 13), we can assume the distributions are NOT normal because one of the
values is below .05.
30

Table 13

Homogeneity of Variance: For this research question, the assumption is that variances are
equal. According to the Shapiro-Wik test for homogeneity of variance, we are provided with a p-value
of .0000000001 (Table 14) between groups, which is less than .05, so we must reject the null
hypothesis that the variances are equal because there is a significant statistical difference. This
assumption has been violated.

Table 14

The tests for normality and homogeneity of variance demonstrate that our assumptions have been
violated as there is not a normal distribution, nor are the variances equal. However, despite this we will
still run a t-test.

Section 4: Run the t-tests (even if the assumptions are not upheld) to answer your hypotheses
The tests for normality and homogeneity of variance demonstrate that our assumptions have
been violated as there is not a normal distribution, nor are the variances equal. However, despite this
we will still run a t-test.
31

Section 5: Present the results of the t-tests using text and tables
If we look at the p-value for the equal variances not assumed (due to the fact that the test for
homogeneity of variance proved the variances were not equal), we see the significance of .000 (Table
15). Because this value is below the demarcation point of .05 we must reject the null hypothesis that
the scores of the students in Dr. Hughes’ class are equal to the scores in Dr. made-Up’s class.

Table 15

Section 6: Interpret the results to answer your research questions


Our research question was are the scores of students in Dr. Hughes’ class equal to the scores of
students in Dr. Made-up’s class? Before we ran our t-test, we checked our assumptions by running a
test of normality and a test of homogeneity of variance. Based on the p-values of .423 and .037 for the
test of normality we concluded that the distributions were not equal, thus we had to reject the
assumption of normal distributions. When we ran the homogeneity of variance test we got the p-value
of .0000000001,which meant we had to reject our assumption of the variances being equal as there is
a significant statistical difference, however we ran the t-test anyways. After running the t-test, we were
given a p-value of .000, which is below the demarcation point of .05. Therefore, we must reject the null
hypothesis that scores of students in Dr. Hughes’ class ARE equal to the scores of students in Dr. Made-
Up’s class and accept the alternative hypothesis that the scores of students in Dr. Hughes’ class are
NOT equal to the scores of students in Dr. Made-Up’s class. . Thus, the answer to our research question
is no, the scores of students in Dr. Hughes’ class are not equal to the scores of students in Dr. Made-
Up’s class.
32

Jessica Buckle
Winter 2020

EDUC 700

Assignment 4, Terminology, Research Questions,


Hypothesis Testing and Assumptions of Parametric
Statistics

1. Draw a picture of a population, a sample,


and explain both. ← POPULATION: a group of individuals that have
shared parameters; All the occurrences with certain
characteristics that you can gain access to; A specific
group that one chooses to study; This is the group you
would like to sample from because this is the group
you are interested in generalizing to

← SAMPLE: a selection from a standardized procedure


that adequately reflects characteristics of the
population; Process of selecting units (i.e. people,
organizations) from a population of interest; Samples
are taken randomly; A smaller group of data, a portion
or a subset of a population

2. What is sampling error?


Sampling error is when your sample does not adequately represent the population. It is
the difference between the sample and population values.
33

3. What is a hypothesis?
A hypothesis is a proposed explanation for a phenomenon. This might be something we
believe, something we are interested in, maybe just a question we have. For it to be a
research hypothesis, the hypothesis must be testable using a statistical analysis that meets
the established criteria for that analysis.
4. What is a null hypothesis?
A null hypothesis is a general statement or default position that there is nothing
significantly different happening, like there is no association among groups or variables,
or that there is no relationship between two or more measured phenomena. For example,
if the research question was: Is the birth rate in California equal to the birth rate in Maine,
the null hypothesis would be: The birth rate in California is equal to the birth rate in
Maine.
5. What is the purpose of a null hypothesis?
The purpose of the null hypothesis is to effectively express a testable hypothesis. It
effectively communicates the common/default belief/position. Research questions can be
confusing, the null hypothesis is meant to be straightforward.
6. What is an example of non-directional research hypothesis?
RQ: Does the birth rate in California equal the birth rate in Maine?
RH: The birth rate in California equals the birth rate in Maine.
7. Explain the non-directional research hypothesis and why it is non-directional.
These are examples of non-directional because it uses the word “equal.” It is a two-tailed
test because you are looking for the 5%, 2.5% on either end of the distribution. It will be
unlikely to use or see a directional test. A non-directional test is more robust and still
provides you with the same information sought after in a directional test.
34

8. What is an example of directional research hypothesis?


RQ: Is the birth rate in California higher than the birth rate in Maine?
RH: The birth rate in California is higher than the birth rate in Maine.
9. Explain the directional research hypothesis and why it is directional.
This is a good example of a directional hypothesis because it uses the words “greater
than.” It is a one-tailed test because you are looking for the 5% on one side.
10. What is a one-tailed test?
A one-tailed test is a statistical test in which the critical area of a distribution is one-sided
so that it is either greater than or less than a certain value, but not both. ... A one-tailed
test is also known as a directional hypothesis or directional test.
If you are using a significance level of .05, a one-tailed test allots all of your alpha to
testing the statistical significance in the one direction of interest. This means that .05 is
in one tail of the distribution of your test statistic. When using a one-tailed test, you are
testing for the possibility of the relationship in one direction and completely disregarding
the possibility of a relationship in the other direction. The one-tailed test provides more
35

power to detect an effect in one direction by not testing the effect in the other direction.

11. What is a two-tailed test?


If you are using a significance level of 0.05, a two-tailed test allots half of your alpha to
testing the statistical significance in one direction and half of your alpha to testing
statistical significance in the other direction. This means that .025 is in each tail of the
distribution of your test statistic. When using a two-tailed test, regardless of the direction
of the relationship you hypothesize, you are testing for the possibility of the relationship
in both directions.

12. What is different about the research and null hypothesis?


A null hypothesis is a hypothesis that says there is no statistical significance between the two
variables. It is usually the hypothesis a researcher or experimenter will try to disprove or
discredit.
36

The null hypothesis states there is no relationship between the measured phenomenon (the
dependent variable) and the independent variable. You do not need to believe that the null
hypothesis is true to test it. On the contrary, you will likely suspect that there is a relationship
between a set of variables. One way to prove that this is the case is to reject the null hypothesis.
Rejecting a hypothesis does not mean an experiment was "bad" or that it didn't produce results.
In fact, it is often one of the first steps toward further inquiry.
Note: In quantitative research, Dr. Hughes believes that it is much easier to write a research
question. I take the Null hypothesis and add in more detail to develop a research question. OR,
I have a question and I consider the best statistical test to answer that question. Then I adjust
my question to align with null hypothesis of that statistical test. Writing good research
questions is difficult. I encourage you to keep a notebook with you so that you can write down
questions you have and then try to find ways to answer those questions. Your approach might
be quantitative, qualitative, or mixed methods...
13. In statistics, what is probability?
Probability is the measure of the likelihood that an event will occur in a Random
Experiment. Probability is quantified as a number between 0 and 1, where, loosely
speaking, 0 indicates impossibility and 1 indicates certainty. The higher the probability of
an event, the more likely it is that the event will occur.

Example: A simple example is the tossing of a fair (unbiased) coin. Since the coin is fair,
the two outcomes (“heads” and “tails”) are both equally probable; the probability of
“heads” equals the probability of “tails”; and since no other outcomes are possible, the
probability of either “heads” or “tails” is 1/2 (which could also be written as 0.5 or 50%).
14. Why is probability important?
What we are doing in class is the basics. But you can use statistics to determine the
impact of something like a treatment. Is statistics only concerned with the middle?
Simple answer NO. I want to know why a group exists or does not exist near what is
considered the “middle”. Quantitative research is a “method” that uses statistics to help
answer a question.
Probability is a notion which we use to deal with uncertainty. If an event can have an
number of outcomes, and we don't know for certain which outcome will occur, we can
use probability to describe the likelihood of each of the possible events. ... There are two
possible outcomes: the coin could come up heads or tails.
15. On a distributions curve, what is an asymptotic tail?
An asymptotic value comes close to but never touches the x-axis. This is significant
because you’re looking for differences (situations where people exist in the tails) and this
makes you aware that those people exist.
37

16. Why is the idea of an asymptotic tail so important?


An asymptotic tail is so important because it makes you aware that people exist that far
out from the median.
17. Describe the Central Limit Theorem and why it is important:
This is the theory that if you take a large enough sample from a population (especially
random), it will be normally distributed.
18. Draw a normal distribution, with standard deviation sections and percentage
distribution of cases at each standard deviation:
38

19. What are standard scores?


Standard scores are a standardized way of interpreting something. It comes from a
distribution with a predefined mean and standard deviation. Examples are the SAT or the GRE to
see the differences. Change scores to scale out of 100.
20. What is a z-score and why is it important? (Note: https://www.simplypsychology.org/z-
table.html)
A z-score is a standardized measure for how far a raw score is from the mean. It’s
important because if tells us how many standard deviations from the mean a single value is. We
are looking for people who exist out of the significance level.

21. Find, copy, and paste a relatively high quality z-score table here (Note: the image
should have a normal distribution pictured or something indicated like “Table Values
39

Represents Area to the left of the Z-score” AND both positive and negative z-score values
(Note: https://www.math.arizona.edu/~rsims/ma464/standardnormaltable.pdf):
40

22. What does having a significance value of .05 and a z-score of 2.6 tell you?
It tells us that that individual person is way out in the extremes because you are looking
at the tail. Only 4.7% of people exist beyond the z-score in this instance. They exist in a zone of
statistical significance/difference.
41

Jessica Buckle
Winter 2020

EDUC 707

Assignment 5, Terminology, Significance, Type I and


II Errors, Confidence Intervals, Effect Size, Power,
Z-Test
Define the following terms and answer the questions:
1. What is statistical significance?
Statistical significance are values in the test that are significantly different. It is
something that is unlikely to happen based on chance alone. It is significant because it
exists in the tails and we want to know why they are different.
2. What is significance level?
The significance level (also refers to the critical value or confidence level) is a value that
you pick, such as .05- you’re looking for values less than or greater than.
3. Why is significance level important?
It is important because it tells us the probability of rejecting the null hypothesis when it is
true. This is also referred to as risk. For example, a significance level of .05 indicates a
5% risk of concluding that a difference exists when there is no actual difference.
4. What does significance level tell you (i.e. the researcher)?
The significance level tells you the probability of rejecting the null hypothesis. The lower
the significance levels indicate that you require stronger evidence before you reject the
null hypothesis. It tells us how confident we are in the hypothesis. For example, we are
95% confident that female and male faculty salaries are the same at the same institution
(example from Assignment 3).
5. What is a type I error?
Claim it’s true but it’s not, false positive. Significance level sets the risk of getting a type
I error. We are claiming what we are measuring is perceived to be true but in reality it is not true
(untrue) or we are accepting the null when it is not true (or vis versa-rejecting the null when it is
true). When you increase your risk of getting a type 1 error you decrease your risk of a type II
error and vis versa, so you need a balance between the two which is why we generally accept .05
because it is in-between. It can only happen when you reject the null: reject the null when in
reality there is no difference. You cannot make a type I error if you accept the null.
42

6. What is a type II error?


If in reality it’s true and we say it’s false it is a type II error, false negative. It can only
happen when you accept the null.
7. Draw a table explaining the possible outcomes of a hypothesis test (Note: from the book):

8. Explain the statement, just because something is statistically significant does not mean it
is practically significant.
Statistically significant means nothing if the study is not designed well or if there is an
insufficient effect size or the power is low, meaning that it is not practically significant (it does
not have a big enough effect for it to make a difference). For example, a person with a doctorate
is not necessarily smarter or a better person than a person without a doctorate.
9. What are the steps/events in making an inference?
(1) Random representative sample; (2) Select appropriate test for statistical significance
(i.e. treatment and control groups); (3) Results (i.e. difference between groups); (4) Conclusion
(i.e. why are the groups different?)
10. What are the steps in the plan for testing significance? (step 2 of making an inference)
1. State the null hypothesis: usually state the research question first
2. Selecting the level of significance, also level of risk for type 1 error
43

3. Select appropriate test


4. Compute test statistic value
5. Determine the critical value needed for rejection
6. Compare obtained and critical values
7 and 8. Less than .05 reject null, greater than .05 accept null (if .05 is your level of
significance)
11. Explain a test statistic:
A test statistic is a statistic used in statistical hypothesis testing. It is a numerical
summary of a data-set that reduces the data to one value that can be used to perform the
hypothesis test. For example, in the research question asking if the height of basketball players
was equal to the height of football players, the heights of the players would be the test statistic.
12. Explain a critical value:
A critical value (also known as the significance level ) refers to the t statistic or other test
statistic and tells us the probability of rejecting the null hypothesis when true.
13. Draw a picture of a normal distribution and label when you accept and reject the null
hypothesis:

14. Explain the confidence interval:


The confidence interval is a range of values so defined that there is a specified probability
that the value of a parameter lies within it. For example, a 95% confidence interval means that
you would be correct 95% of the time. It is the best estimate for a range of population values.
15. What is effect size?
Effect size is a quantitative measure of the magnitude of the experimenter effect. The
larger the effect size the stronger the relationship between two variables. Effect size is a measure
of the strength of the relationship between two variables. In a research setting, it is not only
44

helpful to know whether results have a statistically significant effect, but also the magnitude of
any observed effects.
16. Why is effect size important?
Effect size is important because the larger the effect size the stronger the relationship
between two variables. It helps us to not only know if it is statistically significant but if it is
practically significant.
17. What is statistical power?
Power is the probability of rejecting the null hypothesis when, in fact, it is false. It is the
probability of avoiding a type II error. It is the probability of making a correct decision (i.e. to
reject the null when the null hypothesis is false).
18. Why is statistical power important?
Statistical power is important because it tells us the probability of rejecting the null
hypothesis when, in fact, it is false. It is also important because it helps us avoid a Type II error.
It helps us to discover if there is a statistically significant difference as well as to help us to
determine how big of a sample size we need.
Note: In addition to these Notes you will run a Z-Test using data in Blackboard for
assignment 5.
Research Question 1:
RQ1: Does a sample of male professional basketball players equal the average height of
professional basketball players?
Null Hypothesis 1: The height of the sample population of professional male basketball players
equals the average height of male professional basketball players.
Alternative Hypothesis 1: The height of the sample population of male basketball players does
not equal the average height of male professional basketball players.
Assumptions for RQ1: We cannot check the normality of the population just of our sample
therefore we are unable to run a normality test. We are assuming that the normality and the
homogeneity of variance are accurate as these tests cannot be easily run in SPSS.
Table 1
45

Results for RQ1: The p-value of .01294 is below .05 so the sample mean does not equal the
population. Therefore, we must reject the null hypothesis and accept the alternative hypothesis.
Thus, the answer to RQ1 is the sample population of male professional basketball players does
not equal the average height of male professional basketball players.
Research Question 2:
RQ2: Does the sample populations’ fasting glucose level equal the average glucose level of
nondiabetics’ glucose levels?
Null Hypothesis 2: The sample populations’ fasting glucose level is equal to the average glucose
level of nondiabetics.
Alternative Hypothesis 2: The sample populations’ fasting glucose level is not equal to the
average glucose level of nondiabetics.
Assumptions for RQ2: We cannot check the normality of the population just of our sample
therefore we are unable to run a normality test. We are assuming that the normality and the
homogeneity of variance are accurate as these tests cannot be easily run in SPSS.
Table 2

Results for RQ2: The p-value of .00000 is below .05 so the sample mean does not equal the
population. Therefore, we must reject the null hypothesis and accept the alternative hypothesis.
Thus, the answer to RQ2 is the sample populations’ fasting glucose level is not equal to the
average glucose level of nondiabetics.
Research Question 3:
RQ3: Does the sample percentage of adults with a college degree equal the general percentage
of the population with a college degree?
Null Hypothesis 3: The sample percentage of adults with a college degree is equal to the general
percentage of the population with a college degree.
Alternative Hypothesis 3: The sample percentage of adults with a college degree is not equal to
the general percentage of the population with a college degree.
46

Assumptions for RQ3: We cannot check the normality of the population just of our sample
therefore we are unable to run a normality test. We are assuming that the normality and the
homogeneity of variance are accurate as these tests cannot be easily run in SPSS.
Table 3:

Results for RQ3: The p-value of .00000 is below .05 so the sample mean does not equal the
population. Therefore, we must reject the null hypothesis and accept the alternative hypothesis.
Thus, the answer to RQ3 is the sample percentage of adults with a college degree is not equal to
the general percentage of the population with a college degree.
47

Jessica Buckle
Winter 2020

EDUC 707

Assignment 6, Terminology, ANOVA, F-Statistic


Define the following terms and answer the questions:
1. What is an ANOVA?
An ANOVA is a type of inferential statistic used to determine if there is a significance
between two or more means. It is the analysis of variance; the benefit is that it is used to compare
two or more means. It is more powerful than a t-test because you can introduce more variables
(i.e. checking if gender of the test taker, ethnicity of the test taker, or years of experience of the
teacher impacted scores on a test).
2. What are the different types of ANOVAs?
There are basically two types but there are more. The two primary belong to the category
of Between-Subjects ANOVA. These are used when assessing differences between independent
groups on continuous independent variables. For example, examining the test scores of one class
versus the test scores of another class. There is a One-way ANOVA, which is used when
assessing differences between independent groups on one continuous independent variable (i.e.
interval or ratio data). For example, what is the difference between the time taken to complete
the final in Dr. Hughes, Dr. Madeup, and Dr. Fabricated’s statistics class.
The Factorial ANOVA is used when assessing differences between independent groups
on two or more continuous independent variables (i.e. interval or ratio data). This test is more
popular because you do not have to run the test multiple times and it is more powerful than
others because you can introduce more variables. It helps you to answer more in-depth, detailed
questions. For example, the difference between the time taken to complete the final AND the
grade in Dr. Hughes, Dr. Madeup, and Dr. Fabricated’s statistics class.
3. What are ANOVAs used for?
ANOVAs are used to compare two or more means. For example, checking the test scores
between two difference classes or checking to see if gender affects test scores.
4. What is an F-statistic?
The F statistic is a value you get when you run an ANOVA test to find out if the means
between two or more populations are significantly different. F-statistics for ANOVAs are similar
to the t-statistic for t-tests.
5. Why is the F-statistic important?
48

It is important because it tells you if the means are significantly different. It lets us know
if we need to accept or reject the null hypothesis. This is still based on the significance of
examining if the value is below or above .05.
6. Explain ANOVA: two-factor with replication:
A two-way ANOVA with replication is performed when you have two groups and
individuals within that group are doing more than one thing (i.e. taking two tests). An example
would be if you were comparing students’ tests scores across different tests from two or more
colleges.
7: Explain ANOVA: two-factor without replication:
A two-way ANOVA without replication is if you only have one group. An example
would be if you were comparing one group of students’ scores across different tests.
8. What is a factorial analysis of variance?
A factorial analysis of variance is an analysis with more than one factor or independent
variable. An example would include two factors such as gender (male and female) and treatment
(high impact or low impact exercise program), and the outcome (weight loss).
9. Why is a factorial analysis of variance important?
Factorial analysis of variance is important because it allows us to test for multiple factors
and between two or more groups.
10. What is the main effect in factorial analysis of variance?
The main effect is examining one of the independent variables effect on the dependent
variable.
11. When would you use a factorial ANOVA rather than a simple ANOVA?
You would use a factorial ANOVA over a simple ANOVA when you have multiple
independent variables to compare. For example, if you are looking at does ethnicity or gender
affect test scores you would run a factorial ANOVA because you are looking at the possible
effect of ethnicity and/or gender on test scores.
Note: In addition to these Notes you will run ANOVAs using data in Blackboard for
assignment 6.
Research Question 1:
RQ1: Does a difference in business startup costs exist between Pizza, Bakery, Shoe, Gift, and
Pet shops/stores?
Null RQ1: Mean startup costs are equal for Pizza, Bakery, Shoe, Gift, and Pet shop/stores.
Alternative RQ1: Mean startup costs are NOT equal for Pizza, Bakery, Show, Gift, and Pet
shop/stores.
49

Assumptions for RQ1: To check our assumptions that the data is normally distributed and that
the homogeneity of variance is equal we ran the tests. Based on Table 1, we can see based on the
significance values being below .05 that they are not normally distributed. However, as the
homogeneity of variance is more important, we ran that test anyways. Based on Table 2 we see
the mean significance of .173, which is above the critical value of .05 so we accept the null
hypothesis and will run the One-way ANOVA.

Table 1
50

Table 2

Results for RQ1: If we look at the ANOVA test in Table 2 and look at the significance value
between groups, we get a significance value of .000. This is below the critical value of .05 so we
reject the null hypothesis that all mean start-up costs are equal, and we must accept the
alternative hypothesis that all the means are not equal. Therefore, the answer to RQ1 is: Yes,
there is a difference in business startup costs exist between Pizza, Bakery, Shoe, Gift, and Pet
shops/stores.
Research Question 2:
RQ2: Which business has a higher start-up cost: Pizza, Bakery, Shoe, Gift, and Pet shops/stores?
Null RQ2: The mean startup costs are equal for Pizza, Bakery, Shoe, Gift, and Pet shop/stores.
51

Alternative RQ2: The mean startup costs are NOT equal for Pizza, Bakery, Show, Gift, and Pet
shop/stores.
Assumptions for RQ2: To check our assumptions that the data is normally distributed and that
the homogeneity of variance is equal we ran the tests. Based on Table 1, we can see based on the
significance values being below .05 that they are not normally distributed. However, as the
homogeneity of variance is more important, we ran that test anyways. Based on Table 2 we see
the mean significance of .173, which is above the critical value of .05 so we accept the null
hypothesis and will run the One-way ANOVA.

Table 1
52

Table 2

Results for RQ2: If we look at the ANOVA test in Table 2 and look at the significance value
between groups, we get a significance value of .000. This is below the critical value of .05 so we
reject the null hypothesis that all mean start-up costs are equal, and we must accept the
alternative hypothesis that all the means are not equal. If we go back to the Descriptives in Table
1, we can see the mean for group 1 (Baker) is higher than the other groups. Therefore, the answer
to RQ2 is: There is a higher start-up cost for bakers than there is for pizza, shoe stores, gift
shops, or pet stores.
Research Question 3:
RQ 3: Does day of the week and section of the paper impact number of responses a paper
advertisement received?
53

Null Hypothesis RQ3 #1: Factors 1 and 2 have no interaction OR the magnitude of responses is
equal across all days of the week and sections of the paper
Null Hypothesis RQ3 #2: Factor 1 mean is the same for a groups OR number of responses are
equal based on day advertisement is posted (main effect)
Null Hypothesis RQ3 #3: Factor 2 mean is the same for a groups OR number of responses are
equal based on section of the paper that the advertisement is posted (main effect)
Alternative RQ3 #1: Factors 1 and 2 have interaction, Alternative RQ2 #3: Factor 1 mean is
not equal for a groups, Alternative RQ3 #3: Factor 2 mean is not equal for a groups
Assumptions RQ 3: We need to check our assumptions by running a test for normality and a
test for homogeneity of variance. Based on Table 3, we can see a significance value of .298,
which is above the critical value of .05 so our assumption that there is a normal distribution is
correct. If we look at our Levene’s Test in Table 4 and examine the significance value based on
means we get a value of .931, which is above .05 so we accept the null hypothesis and will run
the Two-way ANOVA.
Table 3

Table 4
54

Results RQ3:
Based on the p-values from Table 5, we can see we must reject all three null hypotheses
and accept all three alternative hypotheses. If we look at Factor 1&2 the p-value is .000, which is
below .05 so we reject null hypothesis RQ2 #1 and accept Alternative RQ2 #1 that there is
interaction between the two variables and that they are not equal across all days of the week and
sections of the paper. If we look at Factor 1, we get a p-value of .000, which is below .05 so we
reject the null hypothesis RQ2 #2 and accept the Alternative RQ2 #2 that the mean of responses
to advertisements is based on the days of the week is not equal. If we look at Factor 2, we have a
value of .000, which is below .05 so we reject Null Hypothesis RQ2 #3 and accept Alternative
RQ2 #3 that the number of responses for advertisements is not equal across sections of the
newspaper. Therefore, the answer to RQ#2 is yes, the day of the week and the section of the
paper impact the number of responses a paper advertisement received.
Table 5

Research Question #4
RQ4: What is the best day to advertise in the paper?
Null Hypothesis RQ4 #: The magnitude of responses is equal across all days of the week.
Alternative RQ4: The magnitude of responses is NOT equal across all days of the week.
Assumptions RQ4: We need to check our assumptions by running a test for normality and a test
for homogeneity of variance. Based on Table 3, we can see a significance value of .298, which is
above the critical value of .05 so our assumption that there is a normal distribution is correct. If
we look at our Levene’s Test in Table 4 and examine the significance value based on means we
get a value of .931, which is above .05 so we accept the null hypothesis and will run the Two-
way ANOVA.
Table 3
55

Table 4

Table 5
56

Table 6

Results RQ4: Based on the p-values from Table 5, we can see we must reject the null
hypotheses and accept the alternative hypotheses. If we look at Factor 1, we get a p-value of
.000, which is below .05 so we reject the null hypothesis RQ4 and accept the Alternative RQ4
that the mean of responses to advertisements is based on the days of the week is not equal. If we
look at the data from Table 6 and examine the significance values and the mean differences we
can tell that the best day to advertise in the paper is Friday. Therefore, the answer to RQ#4 is
Friday is the best day of the week to place an ad to impact the number of responses a paper
advertisement received.
Research Question #5
RQ5: What is the best section of the paper to advertise in?
Null Hypothesis RQ5: The magnitude of responses is equal across all sections of the paper.
Alternative RQ5: The magnitude of responses is NOT equal across all sections of the paper.
Assumptions RQ5: We need to check our assumptions by running a test for normality and a test
for homogeneity of variance. Based on Table 3, we can see a significance value of .298, which is
above the critical value of .05 so our assumption that there is a normal distribution is correct. If
57

we look at our Levene’s Test in Table 4 and examine the significance value based on means we
get a value of .931, which is above .05 so we accept the null hypothesis and will run the Two-
way ANOVA.
Table 3

Table 4

Table 5
58

Table 7

Results RQ5: Based on the p-values from Table 5, we can see we must reject the null
hypotheses and accept the alternative hypotheses. If we look at Factor 2, we get a p-value of
.000, which is below .05 so we reject the null hypothesis RQ5 and accept the Alternative RQ5
that the mean of responses to advertisements is based on the sections of the week is not equal. If
we look at the data from Table 7 and examine the significance values and the mean differences
we can tell that the best section of the paper is the news section. Therefore, the answer to RQ#5
is the news section is the best section to place an ad to impact the number of responses a paper
advertisement received.
Research Question #6
RQ6: What is the best day and section of the newspaper to place an advertisement?
59

Null Hypothesis RQ6 #1: Factors 1 and 2 have no interaction OR the magnitude of responses is
equal across all days of the week and sections of the paper
Null Hypothesis RQ6 #2: Factor 1 mean is the same for a groups OR number of responses are
equal based on day advertisement is posted (main effect)
Null Hypothesis RQ6 #3: Factor 2 mean is the same for a groups OR number of responses are
equal based on section of the paper that the advertisement is posted (main effect)
Alternative RQ6 #1: Factors 1 and 2 have interaction, Alternative RQ2 #6: Factor 1 mean is
not equal for a groups, Alternative RQ6 #3: Factor 2 mean is not equal for a groups
Assumptions RQ6: We need to check our assumptions by running a test for normality and a test
for homogeneity of variance. Based on Table 3, we can see a significance value of .298, which is
above the critical value of .05 so our assumption that there is a normal distribution is correct. If
we look at our Levene’s Test in Table 4 and examine the significance value based on means we
get a value of .931, which is above .05 so we accept the null hypothesis and will run the Two-
way ANOVA.
Table 3

Table 4
60

Table 5

Table 6

Table 7
61

Results RQ6: Based on the p-values from Table 5, we can see we must reject all three null
hypotheses and accept all three alternative hypotheses. If we look at Factor 1&2 the p-value is
.000, which is below .05 so we reject null hypothesis RQ6 #1 and accept Alternative RQ6 #1 that
there is interaction between the two variables and that they are not equal across all days of the
week and sections of the paper. If we look at Factor 1, we get a p-value of .000, which is below
.05 so we reject the null hypothesis RQ6 #2 and accept the Alternative RQ6 #2 that the mean of
responses to advertisements is based on the days of the week is not equal. If we look at Factor 2,
we have a value of .000, which is below .05 so we reject Null Hypothesis RQ6 #3 and accept
Alternative RQ6 #3 that the number of responses for advertisements is not equal across sections
of the newspaper. If we look at the data from Table 6 and examine the significance values and
the mean differences we can tell that the best day to advertise in the paper is Friday. If we look at
the data from Table 7 and examine the significance values and the mean differences we can tell
that the best section of the paper is the news section. Therefore, the answer to RQ#6 is Fridays in
the news section is the best place to place an ad to impact the number of responses a paper
advertisement received.
62

Jessica Buckle
Winter 2020

EDUC 707

Assignment 7, Terminology, Correlations, Reliability,


Validity, and Generalizability
Define the following terms and answer the questions:
1. What is a correlation?
Correlation is how one variable changes as another variable changes. It examines all data
not just one data point and are linear models of relationships. Association NOT causation. Just
because they correlate does not mean it is related.
2. What is a correlation coefficient?
The correlation coefficient is a number that reflects the how the change in one variable
changes another variable.

3. What is the possible range of the correlation coefficient and why?


The correlation coefficient range is from -1 to 1. -1 represents a negative correlation (or
indirect correlation); as one variable increases, the other decreases. A correlation
coefficient of -1 is basically saying that all instances of variable A are equal to the same
negative instances of variable B.
Positive 1 represents a positive correlation or direct correlation, as one variable
increases, the other variable increases. A correlation coefficient of 1 is basically saying
that all instances of variable A are equal to the same instances of variable B.
0 represents no correlation, as one variable increases, the other variable acts in an
indeterminate way.
63

4. What are univariate correlations?


Univariate correlation is a trick term because they do not exist. Uni means one, you need
more than one variable to check correlation.
5. What are bivariate correlations?
All correlations must AT LEAST be bivariate. This means there are two variables. For
example, the more time you study the better you will do on an exam.
6. What are multivariate correlations?
Multivariate correlations are when you are basically looking for correlations between a
dependent variable (i.e. outcome variable) and any combination of two or more independent
variables (i.e. predictive variables). For example, the more time you study and the more sleep
you get the better you will do on an exam.
7. Draw a scatter plot with a positive correlation:
64

8. Draw a scatter plot with a negative correlation:

9. Draw a scatter plot with no correlation:


65

10. Explain a Pearson r correlation:


The Pearson correlation coefficient is also referred to as Pearson’s r, which is the
measure of the linear correlation between two variables X and Y. Used over Kendall and
Spearman correlation in normally distributed data.
11. Explain the assumptions of a Pearson r correlation:
◦ Each variable is measured on a continuous level (i.e. interval or ratio) (Note: If one or
both variables are ordinal in measurement, then a Spearman correlation could be
conducted instead.)
◦ Related pairs (Note: Each participant or observation should have a pair of values. If the
correlation was between weight and height, then each observation used should have both
a weight and a height value.)
◦ Absence of outliers
◦ Normality of Distribution for each variable.
◦ Linearity (i.e. scatterplot forms a straight line)
◦ Homoscedasticity (Note: this is like homogeneity. Homoscedasticity is the distance
between each point and the line describing the data are equal.)
The more assumptions you do not uphold the worse your results will be.
12. What are examples of research questions for a Pearson r correlation?
1. Is there a statistically significant relationship between variable 1 and variable 2?
2. Is there a relationship between variable A and variable b?
3. As variable A increases, does variable B increase?
4. What is the correlation between crime rate and college graduation?
13. Explain the Kendall rank (i.e. Kendall tau (τ)) correlation:
66

The Kendall rank correlation, commonly referred to as Kendall’s coefficient, is a non-


parametric measure of relationships between columns of ranked data (or the strength and
direction of association that exists between two variables measures on at least an ordinal scale).
The Tau correlation coefficient returns a value of 0 to 1 where 0 is no relationship and 1 is a
perfect relationship. (suggested for non-normally distributed data)
14. Explain the Spearman rank (i.e. Spearman rho (ρ)) correlation:
The Spearman rank correlation is a non-parametric test that is used to measure the degree
of association between two variables. (suggested for non-normally distributed data)
15. Explain direct and indirect correlations (Note: consider including a table about direct
and indirect correlations similar to this chapter):
A positive or direct relationship is one in which the two variables (we will generally
call them x and y) move together, that is, they either increase or decrease together. ... In a
negative or indirect relationship, the two variables move in opposite directions, that is, as one
increases, the other decreases.

16. What is reliability?


Reliability is the extent to which the outcomes are consistent when the experiment is
repeated more than once.
17. What is validity?
Validity is the extent to which the instruments that are used in the experiment measure
exactly what you want them to measure.
18. Explain the idea of accuracy (i.e. validity) versus reliability (Note: consider having a
picture and description):
You need both but validity is more important than reliability because validity is making
sure you are measuring exactly what you want to measure while reliability is making sure the
67

outcomes are repeatable. In essence, reliability is being able to hit the same spot each time while
validity is taking a good measure or sample.

19. Explain the idea of accuracy (i.e. validity) versus precision (Note: consider having a
picture and description):
68

Accuracy describes the difference between the measurement and the part’s actual value,
while precision describes the variation you see when you measure the same part repeatedly with
the same device.

20. What is test-retest reliability?


Test-retest reliability is when you want to know whether a test is reliable over time. You
correlate the scores from a test given at Time 1 with the same test given at Time 2. For example,
the Bonzo test of identity formation for adolescents is reliable over time.
21. What is parallel form reliability?
Parallel forms reliability is when you want to know whether several forms of a test are
reliable or equivalent. You correlate the scores from one form of the test with the scores from a
second, different form of the same test of the same content. For example, the two forms of the
Regular Guy test are equivalent to one another and have shown parallel forms reliability.
22. What is internal consistency reliability?
Internal consistency reliability is when you want to know whether the items on a test
assess one, and only one, dimension. To do this you would correlate each individual item score
with the total score. For example, all the items on the SMART Test of Creativity assess the same
construct.
23. What is interrater reliability?
Interrater reliability is when you want to know whether there is consistency in the rating
of some outcome. To do this you would examine the percentage of agreement between raters.
This is very important for qualitative research. For example, the interrater reliability for the best-
69

dressed Foosball player judging was .91, which indicates a high degree of agreement of
agreement among judges.
24. What is content validity?
Content validity is when you want to know whether a sample of items truly reflects an
entire universe of items in a certain topic. To do this you would ask Mr. or Ms. Expert to make a
judgement that the test items reflect the universe of items in the topic being measures. For
example, my weekly quiz is my stats class fairly assesses the chapter’s content.
25. What is criterion validity?
Criterion validity is when you want to know whether test scores are systematically related
to other criteria that indicate the test taker is competent in a certain area. To do this you would
correlate the scores from the test with some other measure that is already valid and assesses the
same set of abilities. For example, performance of the EATS test (of culinary skills) has been
shown to be correlated with being a fine chef 2 years after culinary school.
26. What is construct validity?
Construct validity is when you want to know whether a test measures some underlying
psychological construct. To do this you would correlate the set of test scores with some theorized
outcome that reflects the construct for which the test is being designed. For example, it’s true-
men who participate in body contact and physically dangerous sports score higher on the
TEST(osterone) test of aggression.
27. Explain the idea of measurement error (i.e. observational error):
Measurement error is when there is a difference between the value you got when you
measured (what you observed) and the actual value (true value). For example, a witness to a
crime- witnesses will have different observations about height/weight/clothing color.
28. What is random error?
Random error is an error in measurement caused by factors which vary from one
measurement to another. Such as when Chips Ahoy advertised that their cookies had 1000 chips
in every bag.
29. What is systematic error?
Systematic error is when systematic things impact how accurate something is, for
example a using a ruler to measure something but the ruler is inaccurate therefore everything you
measure is going to be inaccurate.
30. Explain three other potential sources of error:
70

1. Environmental errors: The environmental errors occur due to some external conditions of
the instrument such as pressures, temperature, humidity, or magnetic fields.
2. Observational errors: Types of errors that occur due to wrong observations or reading in
the instruments. For example, reading an energy meter reading incorrectly.
3. Theoretical errors: Caused by simplification of the model system. For example, a theory
states that the temperature of the system surrounding will not change the readings taken
when it actually does, then this factor will begin a source of error in measurement.

31. Explain the idea of observed score versus true score:


Observed score is the actual score on the exam and the true score is the person’s actual
ability. Error is the difference between observed and true scores.
32. Explain generalizability in simple terms:
Generalizability is the extension of research findings from a sample to the population
from which the sample was taken. In the simplest of terms, it is a good research design with an
adequate sample.
Note: In addition to these Notes you will run Correlations using data in Blackboard for
assignment 7.
Research Question 1:
RQ1: What is the correlation between crime rate and college graduation?
Null Hypothesis RQ1: There is NO correlation between crime rate and college graduation.
Alternative RQ1: There is a correlation between crime rate and college graduation.
Assumptions for RQ1: As we can see in Table 1, the p-value is .000, which is below .05, so we
must reject the null hypothesis that the data distribution is normal and accept the alternative
hypothesis that the data is not distributed normally. This tells us we might want to consider a
non-parametric test like Kendall test. Based on Table 2, we can see there are also outliers, which
tells us there is no correlation. However, we are going to continue the test. If we look at Table 3,
we can see that the data is not linear. If we look at Table 4, it does look like there is
homoscedasticity as there are only a few outliers.
71

Table 1

Table 2

Table 3
72

Table 4

Results for RQ1:


73

Table 5

Even though we violated our assumptions by not having a normal distribution and by having
outliers, we ran our test for correlation anyways. If we look at Table 5, we can see a correlation
coefficient of -.026 which tells us they are not correlated. If we look at our p-value, it is .856,
which is over the critical value of .05, so we accept the null hypothesis that there is no
correlation. Therefore, the answer to RQ1 is: There is no correlation between crime rate and
college graduation.
Research Question #2:
RQ#2: Is there a correlation between age and household income in the thousands?
Null Hypothesis RQ2: There is NO correlation between age and household income in the
thousands.
Alternative Hypothesis RQ2: There is a correlation between age and household income in the
thousands.
Assumptions for RQ2: As we can see in Table 6, the p-value is .000, which is below .05, so we
must reject the null hypothesis that the data distribution is normal and accept the alternative
hypothesis that the data is not distributed normally. This tells us we might want to consider a
non-parametric test like Kendall test. Based on Table 7, we can see there are also outliers, which
tells us there is no correlation. However, we are going to continue the test. If we look at Table 8,
we can see that the data is not linear. If we look at Table 9, it does not look like there is
homoscedasticity as there are multiple outliers.
74

Table 6

Table 7

Table 8
75

Table 9
76

Results for RQ2:


Table 10

Even though we violated our assumptions by not having a normal distribution and by having
outliers, we ran our test for correlation anyways. If we look at Table 10, we can see a correlation
77

coefficient of .476 which tells us they are not strongly correlated. If we look at our p-value, it is
.000, which is below the critical value of .05, so we reject the null hypothesis that there is no
correlation and accept the alternative hypothesis that there is a correlation. Therefore, the answer
to RQ2 is: There is a correlation between age and household income in the thousands but
because we violated our assumptions this correlation does not have as much meaning.
Research Question #3:
RQ#3: Is there a correlation between age and credit card debt in the thousands?
Null Hypothesis RQ3: There is NO correlation between age and credit card debt in the
thousands?
Alternative Hypothesis RQ3: There is a correlation between age and credit card debt in the
thousands.
Assumptions for RQ3: As we can see in Table 11, the p-value is .000, which is below .05, so
we must reject the null hypothesis that the data distribution is normal and accept the alternative
hypothesis that the data is not distributed normally. This tells us we might want to consider a
non-parametric test like Kendall test. Based on Table 12, we can see there are also outliers,
which tells us there is no correlation. However, we are going to continue the test. If we look at
Table 13, we can see that the data is not linear. If we look at Table 14, it does not look like there
is homoscedasticity as there are multiple outliers.
Table 11

Table 12
78

Table 13

Table 14
79

Results for RQ3:


Table 15

Even though we violated our assumptions by not having a normal distribution and by having
outliers, we ran our test for correlation anyways. If we look at Table 15, we can see a correlation
coefficient of .279 which tells us they are not strongly correlated. If we look at our p-value, it is
.000, which is below the critical value of .05, so we reject the null hypothesis that there is no
correlation and accept the alternative hypothesis that there is a correlation. Therefore, the answer
80

to RQ3 is: There is a correlation between age and credit card debt in the thousands, but because
we violated our assumptions this correlation does not have as much meaning
Research Question #4:
RQ#4: Is there a correlation between age and debt to income ratio?
Null Hypothesis RQ4: There is NO correlation between age and debt to income ratio.
Alternative Hypothesis RQ4: There is a correlation between age and debt to income ratio.
Assumptions for RQ4: As we can see in Table 16, the p-value is .000, which is below .05, so
we must reject the null hypothesis that the data distribution is normal and accept the alternative
hypothesis that the data is not distributed normally. This tells us we might want to consider a
non-parametric test like Kendall test. Based on Table 17, we can see there are also outliers,
which tells us there is no correlation. However, we are going to continue the test. If we look at
Table 18, we can see that the data is not linear. If we look at Table 19, it looks like there is
homoscedasticity.
Table 16

Table 17
81

Table 18
82

Table 19

Results for RQ4:


Table 20

Even though we violated our assumptions by not having a normal distribution and by having
outliers, we ran our test for correlation anyways. If we look at Table 20, we can see a correlation
coefficient of .008 which tells us they are not correlated. If we look at our p-value, it is .810,
which is below the above the critical value of .05, so we accept the null hypothesis that there is
83

no correlation. Therefore, the answer to RQ4 is: There is no correlation between age and debt to
income ratio.
Research Question #5:
RQ#5: Is there a correlation between age and years with current employer?
Null Hypothesis RQ5: There is NO correlation between age and years with current employer.
Alternative Hypothesis RQ5: There is a correlation between age and years with current
employer.
Assumptions for RQ5: As we can see in Table 21, the p-value is .000, which is below .05, so
we must reject the null hypothesis that the data distribution is normal and accept the alternative
hypothesis that the data is not distributed normally. This tells us we might want to consider a
non-parametric test like Kendall test. Based on Table 22, we can see there are also outliers,
which tells us there is no correlation. However, we are going to continue the test. If we look at
Table 23, we can see that the data is not linear. If we look at Table 24, it looks like there is not
homoscedasticity because there are multiple outliers.
Table 21

Table 22
84

Table 23
85

Table 24

Results for RQ5:


Table 25

Even though we violated our assumptions by not having a normal distribution and by having
outliers, we ran our test for correlation anyways. If we look at Table 25, we can see a correlation
coefficient of .554 which tells us they are not strongly correlated. If we look at our p-value, it is
.000, which is below the above the critical value of .05, so we reject the null hypothesis that there
86

is no correlation and accept the alternative hypothesis. Therefore, the answer to RQ5 is: There is
a correlation between age and years with current employer but because we violated our
assumptions this correlation does not have as much meaning.
87

Jessica Buckle
EDUC 707

Assignment 8, Terminology, Correlations, and Linear


Regression
Define the following terms and answer the questions:
1. Explain the idea of causes and associations (hint: just because two variables are
correlated does not mean they have a causal relationship):
Correlations are associations not causations. Two variables are associated but one
variable does not necessarily cause the other variable. Being associated is not as beneficial or
power as being causation. Researchers often try to overreach and claim causation about a
correlation but the claim does not hold up to scrutiny, you can only say as one variable increases
the other variable increases.
2. What is linear regression?
A linear regression is a predictive analysis that is used to explain the relationship between
a dependent variable and an independent variable.
3. Explain the idea of prediction and why it is important:
Prediction is important because it can help us predict the relationship of something. This
is especially beneficial in areas like finance, sales, inventory, and marketing. For example, if we
know how many bananas are bought each week, we can use that to predict how much inventory
we need to have in stock.
4. Explain the line of best fit in a scatterplot:
This line describes all the data; But there is error in many of the points... (i.e. distance
between the line and data point).

5. Explain the error in prediction of a line of best fit:


88

Linear regression consists of finding the best-fitting straight line through the points.
The best-fitting line is called a regression line. ... The error of prediction for a point is the value
of the point minus the predicted value (the value on the line); The error is the distance from the
point to the line.
6. What is a dependent variable?
A dependent variable is a variable that depends on another (the independent variable). In
an example of a grade based on time studying, the grade would be the dependent variable.
7. What is an independent variable?
In independent variable is a variable that is depended upon by another (the dependent
variable). In an example of a grade based on time studying, time would be the independent
variable.
8. What is an easy way to tell the different between an independent and dependent
variable?
An easy way to tell the difference between an independent variable and a dependent
variable is to look at which one is permanent (independent) and which one changes (dependent).
9. What is the difference between a linear and curvilinear relationship?
When a ratio of change is not constant the correlation is supposed to be linear. In another
term, when the entire points on the scatter diagram start to bend close to a smooth curve, then the
correlation is non-linear (curvilinear).

10. What is the basic mathematical function of a straight line? (Note: label the variables or
write them in words)
𝑦 = 𝑚𝑥 + 𝑏. b is the point where the line crosses the y-axis m is the slope of the line x and y are
points on the coordinated system

11. Based on the equation above, what do you need to know to be able to draw a line?
89

We need to know two points through which the line passes.


12. Explain the error of estimate:
Error of estimate is basically the difference between the actual value and the predicted
value. The difference between one point and the line. Not as helpful as the standard error or
estimate.
13. Explain the standard error of estimate:
Standard Error of Estimate is basically the accuracy of the predictions. It is standardized
by taking the average difference between all points and the line.
14. What is a multiple regression?
A multiple regression is a predictive analysis that is used to explain the relationship
between a dependent variable and more than one independent variable.
15. What are the two guidelines for selecting/using predictor variables?
The two guidelines for selecting/using predictor variables is that you shouldn’t use too
many (be selective) and do not be too redundant.
16. Explain the trade-offs associated with number of predictor variables:
More predictors are not always better. In regression, each of the independent variables
add something to the predictive ability. However, if the amount the independent variable adds is
negligible, it might be better not to include it in the model.
Note: In addition to these Notes you will run Regressions using data in Blackboard for
assignment 8.
RQ#1
RQ#1: How does high school dropout rate affect crime rate?
RQ #1 Null Hypothesis: There is no significant relationship between high school drop
out rate and crime rate.
RQ #1 Alternative Hypothesis: There is a statistically significant relationship between
high school dropout rate and crime rate.
RQ #1 Assumptions: Before we can run our regression, we need to check the following
assumptions: linear relationship, multivariate normality, multicollinearity, and homoscedasticity.
Based on the p-value of .111 in Table 1 we can accept the null hypothesis that all independent
variables are normally distributed. Looking at Table 2 we can see that it appears that the data can
be described by a line, but we will run a correlation to verify. Based on the correlation co-
efficient in Table 3, we can see that there is a statistical significance, which tells us we can
accept the null hypothesis that all values exist on a line, even though there might be some error.
Since we only have one independent variable, we do not have to worry about checking
90

multicollinearity. Lastly, Table 4 shows us there is homoscedasticity as most of the data is equal
distance from the line and there are no major outliers, therefore we accept the null that all are
equal distance from the line. We can accept that our assumptions are true and move on to
running our regression.
Table 1

Table 2

Table 3
91

RQ #1 Results:
Table 4

Based on Table 4 we can see there is a statistical significance because .022 is above .05.
Our R squared value is .104, which tells us there is a predictive ability because it is above 0.
Therefore, we must accept the alternative hypothesis that there is a statistically significant
relationship between high school dropout rate and crime rate. And the answer to our research
question is that as high school drop out rate increases it will have a 10% increase impact on the
crime rate.

RQ#2
RQ#2: How does high school dropout rate, reported violent crime rate, annual police
funding, rate of people with a high school degree, rate of people in college, and rate of college
graduates affect crime rate?
RQ #2 Null Hypothesis: Each added independent variable does not improve the fit.
RQ #2 Alternative Hypothesis: Each added independent variable does improve the fit
RQ #2 Assumptions: Before we can run our regression, we need to check the following
assumptions: linear relationship, multivariate normality, multicollinearity, and homoscedasticity.
Based on Table 5, we can see that not all values are above .05 so we must reject the null that it is
normally distributed since not ALL the data is normally distributed. To check linearity we ran a
correlation, shown in Table 6. The correlation matrix in Table 6 shows that half of the variables
are above .05 so we are going to accept the null that all values exist on a line. However, Table 6
also tells us that we cannot accept the null for multicollinearity because the independent
variables seem to be correlated to one another and not independent. However, we will run the
regression anyways. Table 7 shows us that we can accept the null for homoscedasticity as there
are not too many extreme outliers.
92

Table 5

Table 6

Table 7
93

RQ #2 Results:
Table 8

Based on Table 8, our R squared value is .613, which tells us that it is strongly correlated
and a pretty good predictor. Knowing these factors, we can answer our research question by
saying we can say that as the independent variables change, they will influence crime rate by
61%.
RQ#3
RQ#3: Does age impact car accidents?
RQ #3 Null Hypothesis: There is no significant relationship between age and car
accidents.
RQ #3 Alternative Hypothesis: There is a statistically significant relationship between
age and car accidents.
RQ #3 Assumptions: Before we can run our regression, we need to check the following
assumptions: linear relationship, multivariate normality, multicollinearity, and homoscedasticity.
94

Based on a significance value of .488 we can accept the null that the data is normally distributed.
Based on Table 10 it does not look like there is a linear relationship but if we run a correlation,
as seen in Table 11, we can see that there is a correlation so we will accept the null that all the
data exists on a line. Since there is only one independent variable, we do not have to check the
multicollinearity. Based on Table 12 we can accept the null for homoscedasticity as all points
appear equal distance from the line so we will run our regression.
Table 9

Table 10

Table 11
95

Table 12

RQ #3 Results:
Table 13
96

Based on Table 13 we are given an R squared value of .002. Even though this is slight it
is above 0, which means we must reject the null hypothesis and accept the alternative hypothesis
that there is a statistically significant relationship between age and car accidents. To answer our
research question, we can attribute a 2% variability in car accidents based on age.
RQ#4
RQ#4: Does age impact credit card debt?
RQ #4 Null Hypothesis: There is no significant relationship between age and credit card
debt.
RQ #4 Alternative Hypothesis: There is a statistically significant relationship between
age and credit card debt.
RQ #4 Assumptions: Before we can run our regression, we need to check the following
assumptions: linear relationship, multivariate normality, multicollinearity, and homoscedasticity.
Based on a significance value below .05 as seen in Table 14 we can tell that the data is not
normally distributed, and we must reject the null. Based on Table 15, it does not look like there
is a linear relationship so we will run a correlation to check. In Table 16 we see a value of .279
and a significance value of .000, which tells us there is some correlation, but it is not strong.
Because there is only one independent variable, we do not have to check the multicollinearity.
Table 17 shows us the data appears to be equal distance from the line so we will accept the null
for homoscedasticity. Despite not all our assumptions being met, we will run the regression.
Table 14

Table 15
97

Table 16

Table 17
98

RQ #4 Results:
Table 18

According to Table 18, after running our regression, we get an R value of .279 and an R
squared value of .078, which tells us there is a predictive ability because it is above 0. Therefore,
we must accept the alternative hypothesis that there is a statistically significant relationship
between age and credit card debt. And the answer to our research question is that there is almost
an 8% variability in credit card debt based on age.

RQ#5
RQ#5: Does age, years with current employer, years at current address, household
income, debt to income ratio, and other debt impact credit card debt?
RQ #5 Null Hypothesis: Each added independent variable does not improve the fit.
RQ #5 Alternative Hypothesis: Each added independent variable does improve the fit.
RQ #5 Assumptions: Before we can run our regression, we need to check the following
assumptions: linear relationship, multivariate normality, multicollinearity, and homoscedasticity.
Based on Table 19, we can see that we must reject the null and accept the alternative that the data
is not normally distributed. To check for linear relationship, we ran a correlation. Based on Table
20, we can tell from our correlation matrix that there appears to be a linear relationship. We can
also tell from Table 20 that we cannot uphold the null for multicollinearity because it appears
that most of the variables are correlated, however we will still run our regression. According to
Table 21, it looks like we can accept the null for homoscedasticity as it appears most of the data
is the same distance from the line with a few outliers.
Table 19
99

Table 20

Table 21
100

RQ #5 Results:
Table 22

Based on Table 22, we get an R value of .771 and an R squared value of .594. This tells
us that it is strongly correlated and a good predictor. Therefore, we must reject the null
hypothesis and accept the alternative hypothesis. Thus, to answer out research question, knowing
these factors, we can predict that as the independent variables change, credit card debt will vary
by 59%.
RQ#6
RQ#6: Does age affect a violent first crime?
RQ #6 Null Hypothesis: There is no significant relationship between age and a violent
first crime.
RQ #6 Alternative Hypothesis: There is a statistically significant relationship between
age and a violent first crime.
101

RQ #6 Assumptions: Before we can run our regression, we need to check the following
assumptions: linear relationship, multivariate normality, multicollinearity, and homoscedasticity.
Based on a p-value of .000, we can reject the null hypothesis and accept the alternative that the
data is not normally distributed. Based on Table 24 it does not look like there is a linear
relationship, but we will run a correlation to check. Based on Table 25 it looks like there is a
slight correlation so we will accept the null hypothesis that the values exist on a line. Since there
is only one independent variable, we do not have to check multicollinearity. Based on Table 26 it
appears we can accept the null for homoscedasticity as all points seem equal distance from the
line.
Table 23

Table 24

Table 25
102

Table 26

RQ #6 Results:
Table 27

Based on Table 27, we get an R value of .010 and an R squared value that is close to 0,
therefore we will accept the null hypothesis that there is not a significant statistical relationship
between age and a violent first crime. Therefore, the answer to our research question is that age
does not significantly impact violent first crime.
103

Jessica Buckle
Winter 2020

EDUC 707

Assignment 9, Terminology, Nonparametric


Statistics, and Chi-Square
Define the following terms and answer the questions:
Note: I almost never use chi-square, but it is arguably one of the most common nonparametric
statistics. Chi-square is likely the most commonly used statistic for probability. More
specifically chi-square describes a “special” or “unique” distribution indicating probability.
1. Explain how the assumptions of parametric statistics varies from nonparametric
statistics:
The assumptions for a parametric statistic are that there is normal distribution,
homogeneity of variance, and a large enough sample. The assumptions for nonparametric
statistics are that there is non-normal distribution and there is homogeneity of variance. The only
major difference between the two is the distribution assumption.
Note: In the book it says that it requires about 30 participants to approximate a normal
distribution. This is something that is debated between statisticians. I believe that the more
commonly accepted number of participants to approximate a normal distribution is about 32
participants. However, that is no guarantee and the assumption of normal distribution will
always be verified when using parametric statistics.
2. Develop a table describing the test name, when it is used, and sample research questions
(Note: similar to the table in this chapter):
Name of Test When it is Used Sample Research Question
McNemar test for To examine “before and How effective is a phone call
significance of changes after” changes to undecided voters in
influencing them to vote a
certain way on an issue?
Fisher’s exact test To compute the exact What is the exact likelihood
probability of outcomes in a 2 of getting six heads on a toss
X 2 table of six coins?
Chi-square one sample test To determine whether the Did Fruities, Whammies, and
number of occurrences across Zippies each sell an equal
categories is random number of units during a
recent sale?
Kolmogrov-Smirnov test To see whether scores from a How representative is a set of
sample came from a specified judgements about certain
population children at an elementary
104

school of the entire student


body of that school?
The sign test, or median test To compare the medians from Is the median income of
two samples people who voted for
Candidate A greater than the
median income of people
who voted for Candidate B?
Wilcox Rank sum Test To compare the magnitude as Is preschool twice as
well as the direction of effective as no preschool
differences between two experience for helping
groups develop children’s language
skills?
Mann-Whitney U test To compare two independent Did the transfer of learning,
samples measured by number correct,
occur faster for Group A than
for Group B?
Spearman correlation To compute the correlation What is the correlation
between ranks between class rank at the end
of the senior year of high
school and class rank at the
end of the freshman year of
college?
Kruskal Wallis Test To compare the overall How do rankings of
difference between two or supervisors differ among four
more independent samples regional offices?
Friedman two-way analysis To compare the overall How do rankings of
of variance difference between two or supervisors differ as a
more independent samples on function of regional office
more than one dimension and gender?

3. What is a one-sample chi-square?


105

Chi-square is used to determine whether your observed frequencies is what you would
expect. A one-sample chi-square only has one categorical variable. For example, are the number
of respondents equally distributed across all three levels of education (no college, some college,
and college degree)?
4. What is a two-sample chi-square?
A two-sample chi square has two categorical variables, for example it might be used to
test whether preference for school vouchers is independent of political affiliation and gender.
5. What are the eight steps in the chi-square test?
1. State the null and research hypotheses
a. the null hypothesis states that there is no difference in the frequency or the
proportion of occurrences in each category
2. Set the level of risk (or the level of significance of Type I error) associates with the null
hypothesis
a. The Type I error rate is set at .05
3. Select the appropriate test statistic
a. Use the flowchart to determine
4. Compute the test statistic value (called the obtained value)
5. Determine the value needed for rejection of the null hypothesis using the appropriate
table of critical values for the particular statistic
a. Look at degrees of freedom
i. Use this number and the level of risk to look up the critical value on the
chi-square table
6. Compare the obtained value with the critical value
7. And 8. Decision time!
Note: chi-squared is a special case of the gamma distribution. Researching the gamma
distribution will help you explore the chi-square distribution.
Note: In addition to these Notes you will run Nonparametric tests using data in Blackboard
for assignment 9.
RQ1-(Chi-sqaure)
RQ #1: Is there an association between major and socioeconomic status?
Null Hypothesis RQ #1: Major and socioeconomic status are independent.
Alternative Hypothesis RQ #1: Major and socioeconomic status are NOT independent.
Assumptions RQ #1: For a chi-square test the assumptions are that there is a random
sample, a large enough sample, that the observations are independent, and that the observations
are not normally distributes. Based on Table 1 we can accept the null for the assumptions that
there is a random sample and that the sample is large enough as there is a sample of 272
participants. After running a correlation, we can tell based on Table 2 that the variables are not
106

strongly correlated so we will accept the null that the observations are independent. Lastly, we
checked distribution, as seen in Table 3 and 4, and can accept the null that both observations are
not normally distributed.
Table 1

Table 2

Table 3

Table 4
107

Results RQ #1:
Table 5

Table 6

Table 7
108

As you can see in Table 7, there is a disproportionate number of students between


categories of major and socioeconomic status. In order to accept the null hypothesis, we would
need to see the majors equally dispersed among the socioeconomic status groups. However,
since we do not, we must reject the null hypothesis and accept the alternative hypothesis that
major and socioeconomic status are not independent in some populations. Therefore, the answer
to our research question is yes, there is an association between major and socioeconomic status.
RQ2- (Wilcox Sign Rank Sum Test)
RQ #2: Do male faculty salaries equal female faculty salaries?
Null Hypothesis RQ #2: Median male faculty salaries are equal to median female
faculty salaries.
Alternative Hypothesis RQ #2: Median male faculty salaries are NOT equal to median
female faculty salaries.
Assumptions RQ #2: The assumptions for the Wilcox Sign Rank Sum Test are that the
two samples are dependent, that they are independently drawn, that the dependent variable is at
the interval or ratio level, that the independent variable is at least at the ordinal level, and that
there is a homogeneity of variance. We can accept the null for the assumptions that the samples
are dependent, that they are independently drawn, that the dependent variable is at the interval or
ratio level, and that the independent variable is at least at the ordinal scale by looking at the data.
Next we check homogeneity of variance by running a one-way ANOVA as seen in Table 8.
109

Based on a significance value of below .05, we must reject the null and accept the alternative that
the variance between groups is not normal.
Table 8

Results RQ #2:
Table 9
110

As we can see in Table 9, we get an Asymp. Sig value of .330, which is above .05, so we
must accept the null hypothesis that the distributions are similar. Therefore, we must accept the
null hypothesis that median male faculty salaries are equal to median female faculty salaries. So,
the answer to our research question is yes, male faculty salaries equal female faculty salaries.
Although we must take this with a grain of salt since we violated the homogeneity variance
assumption.
RQ3- (Wilcox Sign Rank Sum Test)
RQ #3: Is the percentage of voters who voted Democrat equal to the percentage of voters
who voted Republican in a recent national election?
Null Hypothesis RQ #3: The median of Democratic voters is equal to the median of
Republican voters.
Alternative Hypothesis RQ #3: The median of Democratic voters is NOT equal to the
median of Republican voters.
Assumptions RQ #3: The assumptions for the Wilcox Sign Rank Sum Test are that the
two samples are dependent, that they are independently drawn, that the dependent variable is at
the interval or ratio level, that the independent variable is at least at the ordinal level, and that
there is a homogeneity of variance. . We can accept the null for the assumptions that the samples
are dependent, that they are independently drawn, that the dependent variable is at the interval or
ratio level, and that the independent variable is at least at the ordinal scale by looking at the data.
Next we check homogeneity of variance by running a one-way ANOVA as seen in Table 10.
Levene’s test of homogeneity could not be computed so we must reject the null that the variances
are equal.
Table 10
111

Results RQ #3:
Table 11

As we can see in Table 11, we get an Asymp. Sig value of .850, which is above .05, so
we must accept the null hypothesis that the distributions are similar. Therefore, we must accept
the null hypothesis that median of Democratic voters is equal to the median of Republican
voters. So, the answer to our research question is yes, the percentage of voters who voted
Democrat equal to the percentage of voters who voted Republican in a recent national election.
Although we must take this with a grain of salt since we violated the homogeneity variance
assumption.
RQ4- (Kruskal Wallis)
RQ #4: Does a difference in business startup costs exist between Pizza, Bakery, Shoe,
Gift, and Pet shop/stores?
Null Hypothesis RQ #4: Mean startup costs are equal for Pizza, Bakery, Shoe, Gift, and
Pet shop/stores.
Alternative Hypothesis RQ #4: Mean startup costs are NOT equal for Pizza, Bakery,
Shoe, Gift, and Pet shop/stores.
Assumptions RQ #4: The assumptions for a Kruskal-Wallis test are that there is a
random sample, that the dependent variable is at an ordinal level or above, that the observed
values are independent, and that there is homogeneity of variance. All these assumptions other
than the homogeneity of variance can be verified practically by looking at the data collected. We
can accept the assumption based on the data collected our sample is random. We can also accept
that the dependent variables are at an ordinal level or above. We can also accept that the values
are independent. So, we just need to run the homogeneity of variance.
112

Results RQ #4:
RQ5- (Kruskal Wallis)
RQ #5:
Null Hypothesis RQ #5:
Alternative Hypothesis RQ #5:
Assumptions RQ #5: The assumptions for a Kruskal-Wallis test are that there is a
random sample, that the dependent variable is at an ordinal level or above, that the observed
values are independent, and that there is homogeneity of variance. All of these assumptions other
than the homogeneity of variance can be verified practically by looking at the data collected.
Results RQ #5:

You might also like