Professional Documents
Culture Documents
Unit Iv
Unit Iv
Tests Of Significance
2. Alternative Hypothesis ((Ha)): The alternative hypothesis is the claim that the
researcher wants to test. It represents the assertion that there is indeed an effect
or a difference in the population. The alternative hypothesis can take different
forms depending on the research question, such as stating that the population
parameter is not equal to a specific value, is greater than a specific value, or is
less than a specific value. Using the example of testing a new drug, the
alternative hypothesis might state that the group receiving the drug shows
significantly better outcomes compared to the group receiving a placebo.
● In hypothesis testing, the researcher collects sample data and calculates a test statistic
based on the data.
● The test statistic provides a measure of how much the sample data deviate from what
would be expected under the null hypothesis.
● The conclusion of the hypothesis test depends on whether the observed data provide
sufficient evidence to reject the null hypothesis in favor of the alternative hypothesis, or if
there is not enough evidence to do so.
● First, the researcher formulates two competing hypotheses: the null hypothesis ((H0))
and the alternative hypothesis ((Ha)).
● The null hypothesis represents the status quo or the default assumption, often stating
that there is no effect or no difference in the population.
● The alternative hypothesis, on the other hand, represents the claim the researcher wants
to test, suggesting that there is indeed an effect or a difference.
● Next, the researcher selects an appropriate statistical test based on the type of data and
the research question. Common statistical tests include t-tests, chi-square tests,
ANOVA, and correlation tests, among others.
● After selecting the test, the researcher collects sample data and calculates a test statistic
based on the data. This test statistic provides a measure of how much the sample data
deviate from what would be expected under the null hypothesis.
● Once the test statistic is calculated, the researcher determines the probability of
observing such extreme results if the null hypothesis were true.
● This probability is known as the p-value.
● A low p-value indicates strong evidence against the null hypothesis, while a high p-value
suggests that the null hypothesis cannot be rejected.
● Finally, the researcher compares the p-value to a predetermined significance level (often
denoted as (alpha)), typically set at 0.05. If the p-value is less than or equal to the
significance level, the researcher rejects the null hypothesis in favor of the alternative
hypothesis. If the p-value is greater than the significance level, the null hypothesis is not
rejected.
● Steps:
●
1. State the Hypotheses:
- Null Hypothesis ((H0)): Start by stating the null hypothesis, which represents the
default assumption or the claim that there is no effect or no difference in the population.
This hypothesis is typically denoted as (H_0) and often states that a population
parameter is equal to a specific value or follows a certain distribution.
- Alternative Hypothesis ((Ha)): Next, state the alternative hypothesis, which is the
claim that you are testing. This hypothesis represents the assertion that there is indeed
an effect or a difference in the population. The alternative hypothesis can take different
forms depending on the research question and is denoted as (Ha).
3. Select the Appropriate Test: Choose a statistical test based on the type of data and
the research question being investigated. Common statistical tests include t-tests,
chi-square tests, ANOVA, and correlation tests, among others.
4. Collect Data and Calculate Test Statistic: Collect sample data and calculate a test
statistic based on the data. The test statistic provides a measure of how much the
sample data deviate from what would be expected under the null hypothesis.
5. Determine the Critical Region: Determine the critical region, which is the range of
values of the test statistic that would lead to rejection of the null hypothesis at the
chosen significance level. This critical region is determined based on the selected
significance level and the distribution of the test statistic.
6. Calculate the p-value: Calculate the p-value, which is the probability of observing
results as extreme as or more extreme than the observed data, assuming the null
hypothesis is true. The p-value provides a measure of the strength of evidence against
the null hypothesis.
7. Make a Decision: Compare the p-value to the significance level ((alpha)). If the
p-value is less than or equal to (alpha), reject the null hypothesis in favor of the
alternative hypothesis. If the p-value is greater than (alpha), fail to reject the null
hypothesis.
8. Interpretation: Finally, interpret the results of the hypothesis test in the context of the
research question. If the null hypothesis is rejected, conclude that there is evidence to
support the alternative hypothesis. If the null hypothesis is not rejected, conclude that
there is not enough evidence to support the alternative hypothesis.
● Type I and Type II errors are fundamental concepts in hypothesis testing, representing
the two potential mistakes that researchers can make when interpreting the results of
their analyses.
● Type I errors, often referred to as false positives, occur when the null hypothesis is
incorrectly rejected, indicating an effect or difference in the population when, in reality,
there is none.
● This error is akin to seeing an effect that doesn't actually exist.
● The probability of committing a Type I error is determined by the significance level
((alpha)) chosen for the test. For instance, if the significance level is set at 0.05, there is
a 5% chance of making a Type I error.
● An example of a Type I error in a medical context could be mistakenly concluding that a
new drug is effective in treating a condition when, in fact, it has no therapeutic benefit.
● On the other hand, Type II errors, also known as false negatives, occur when the null
hypothesis is incorrectly not rejected, failing to detect an effect or difference in the
population when one actually exists.
● This error represents a missed opportunity to identify a true effect.
● Unlike Type I errors, the probability of committing a Type II error is influenced by factors
such as sample size, effect size, and the chosen significance level.
● For example, in a medical study, a Type II error might occur if researchers fail to
conclude that a new drug is effective in treating a condition when, in reality, it does have
a beneficial effect.
● This error could have serious implications, as it may result in the rejection of a potentially
beneficial treatment.
● A confidence interval is a range of values that is likely to contain the true value of a
population parameter, such as a mean or proportion, with a specified level of confidence.
● For example, a 95% confidence interval for a mean represents the range of values within
which we are 95% confident that the true population mean lies.
● The width of the confidence interval depends on factors such as the sample size and
variability of the data.
● A narrower confidence interval indicates greater precision in estimating the population
parameter.
● While the p-value provides information about the statistical significance of an effect or
difference, the confidence interval provides information about the precision of the
estimate of a population parameter.
● Together, these measures help researchers interpret the results of statistical analyses,
assess the reliability of study findings, and make informed decisions in medical and
biological research.
● The p-value quantifies the strength of evidence against the null hypothesis.
● It measures the probability of obtaining results as extreme as the observed data,
assuming that the null hypothesis is true.
● A small p-value (typically less than the chosen significance level, e.g., 0.05) indicates
strong evidence against the null hypothesis, suggesting that the observed results are
unlikely to have occurred by random chance alone.
● However, it's important to note that a small p-value does not necessarily imply the
practical significance or importance of the observed effect.
● It only indicates the statistical significance.
● The p-value is influenced by factors such as the sample size, effect size, and variability
of the data.
● Larger sample sizes tend to result in smaller p-values for the same effect size.
Interpretation of the p-value should consider the context of the study and the research
question being investigated.
● It is a measure of statistical significance, not necessarily of clinical or practical
significance.
● A confidence interval provides a range of values within which we are confident that the
true value of a population parameter lies, based on the sample data.
● Commonly used confidence levels include 90%, 95%, and 99%. For example, a 95%
confidence interval for a mean indicates that there is a 95% probability that the true
population mean falls within the interval.
● A narrower confidence interval indicates greater precision in estimating the population
parameter, while a wider interval indicates greater uncertainty.
● The width of the confidence interval depends on factors such as the sample size and
variability of the data. Larger sample sizes generally result in narrower confidence
intervals.
● Confidence intervals provide valuable information about the precision of parameter
estimates and help researchers assess the reliability of study findings. They also allow
for comparisons between different groups or populations.
1. Formulate Hypotheses: Start by stating your null hypothesis (H0) and alternative
hypothesis (Ha).
2. Choose the Type of t-test: Decide which type of t-test is appropriate for your data. The
most common types are:
● Independent samples t-test: Compares the means of two independent groups.
● Paired samples t-test: Compares the means of two related groups.
3. Collect Data: Collect data from your samples or groups.
4. Calculate the Test Statistic: Calculate the t-value using the appropriate formula based
on the type of t-test you're conducting.
● Two t test:
5. Determine the Degrees of Freedom: Calculate the degrees of freedom (df), which
depend on the sample sizes and the type of t-test.
● For one t test: n–1
● For two t test: Welch Satterthwaite formula:
6. Find the Critical Value or P-value: Use a t-table or statistical software to find the critical
value corresponding to your chosen significance level (alpha) or directly obtain the
p-value.
7. Make a Decision: Compare the calculated t-value with the critical value or compare
the p-value with the chosen significance level. If the calculated t-value is greater than the
critical value or if the p-value is less than alpha, reject the null hypothesis. Otherwise, fail
to reject the null hypothesis.
8. Interpret Results: Interpret the findings in the context of your research question or
hypothesis. If you reject the null hypothesis, it suggests that there is a significant
difference between the groups you're comparing.
9. Report Results: Report the results of your t-test, including the calculated t-value,
degrees of freedom, critical value or p-value, and your decision regarding the null
hypothesis.
Paired t-test
● A paired t-test, also known as a dependent t-test or a paired samples t-test.
● It is a statistical test used to compare the means of two related groups.
● This test is appropriate when the same subjects have been measured under two
different conditions or at two different points in time, resulting in pairs of observations.
Steps:
1. Hypotheses: The null hypothesis (H0) typically states that there is no difference
between the means of the two groups, while the alternative hypothesis (H1) states that
there is a difference.
2. Assumptions: The data should be continuous, approximately normally distributed, and
have a similar variance between the two groups.
3. Procedure:
● Calculate the differences between paired observations (e.g., post-test score minus
pre-test score).
● Calculate the mean and standard deviation of the differences.
● Determine the degrees of freedom, typically n-1 where n is the number of pairs.
● Finally, compare the obtained t statistic to a critical value from the t-distribution or use
software to determine the p-value.
4. Interpretation: If the calculated p-value is less than the chosen significance level (e.g.,
0.05), then the null hypothesis is rejected, and it is concluded that there is a significant
difference between the means of the two groups.
2. Set the significance level (α): This is typically set at 0.05, but it can vary depending on the
context and requirements of your analysis.
3. Collect data: Obtain data in the form of a contingency table. A contingency table is a tabular
representation of the frequencies of different categories for two or more variables.
4. Calculate expected frequencies: Compute the expected frequencies under the assumption
that there is no association between the variables. This is done using the formula:
5. Calculate the chi-square statistic: Compute the chi-square statistic using the formula:
6. Determine degrees of freedom: Degrees of freedom ((df)) is calculated as:
7. Find the critical value: Look up the critical value of chi-square in the chi-square distribution
table with (df) degrees of freedom and at the chosen significance level (α).
8. Compare calculated chi-square statistic with critical value: If the calculated chi-square statistic
is greater than the critical value, then reject the null hypothesis. If it is less than the critical value,
fail to reject the null hypothesis.
9. Interpret results: Based on the comparison, make a decision regarding the null hypothesis. If
you reject the null hypothesis, it suggests that there is a significant association between the
categorical variables.
10. Conclusion: Provide a conclusion based on the results of the test, indicating whether there is
evidence to support the alternative hypothesis.
TEST OF PROPORTION
● A test of proportion is a statistical test used to determine whether the proportion of
successes in a sample is significantly different from a hypothesized population
proportion. It's commonly used when dealing with categorical data with two outcomes
(success or failure).
1. Formulate hypotheses:
● Null Hypothesis (H0): The population proportion is equal to a specified value.
● Alternative Hypothesis (H1): The population proportion is not equal to the specified
value.
2. Set the significance level (α): Typically set at 0.05 or another predetermined value.
3. Collect data: Obtain a sample that includes the number of successes and the total
sample size.
4. Calculate the sample proportion: Compute the proportion of successes in the sample
by dividing the number of successes by the total sample size.
5. Calculate the test statistic: For large sample sizes (typically n > 30), you can use the z-test for
proportions. The test statistic (z-score) is calculated as:
6. Determine critical value or p-value: Based on the chosen significance level (α), determine the
critical value from the standard normal distribution table for a two-tailed test or use statistical
software to find the p-value associated with the test statistic.
8. Interpret results: If the null hypothesis is rejected, conclude that there is evidence to suggest
that the sample proportion differs significantly from the hypothesized population proportion.
9. Conclusion: Provide a conclusion based on the results of the test, indicating whether there is
evidence to support the alternative hypothesis.
ANOVA TEST (One Way Analysis of Variance)
● ANOVA (Analysis of Variance) is a statistical test used to determine whether there are
statistically significant differences between the means of three or more independent
groups.
● It is an extension of the t-test for comparing the means of two groups to situations where
there are more than two groups.
● Certainly! One-way Analysis of Variance (ANOVA) is a statistical method used to
compare the means of three or more independent groups.
● It's called "one-way" because there's only one independent variable or factor being
considered.
● ANOVA helps researchers determine whether there are statistically significant
differences between the means of the groups, beyond what would be expected by
random chance.
● In a one-way ANOVA, the data are organized into groups based on the levels of the
independent variable.
● The ANOVA test compares the variation between the group means (the "between-group"
variation) to the variation within each group (the "within-group" variation).
● If the between-group variation is larger than the within-group variation, it suggests that
there are systematic differences between the groups.
● One-way ANOVA has several advantages. It allows researchers to simultaneously
compare multiple groups, which can be more efficient than conducting pairwise
comparisons.
● Additionally, ANOVA provides a way to control for Type I error inflation that can occur
with multiple t-tests.
● However, there are some assumptions that must be met for ANOVA to produce valid
results.
● These include the assumption of normality (the data within each group should be
approximately normally distributed), homogeneity of variances (the variances within each
group should be approximately equal), and independence of observations.
● If the assumptions are met and the ANOVA yields a significant result, post-hoc tests can
be conducted to determine which specific group means differ from each other. Common
post-hoc tests include Tukey's HSD (Honestly Significant Difference), Bonferroni
correction, and Scheffé's method.
● In conclusion, one-way ANOVA is a powerful statistical tool for comparing means across
multiple groups, providing researchers with valuable insights into differences between
experimental conditions or groups.
● Steps:
1. Formulate hypotheses:
- Null Hypothesis (H0): The means of all groups are equal.
- Alternative Hypothesis (H1): At least one group mean is different from the others.
2. Set the significance level (α): Typically set at 0.05 or another predetermined value.
3. Collect data: Obtain data from three or more independent groups. Each group should have
data that represent the variable of interest (e.g., test scores, reaction times, etc.).
4. Check assumptions:
- Independence: Observations within each group should be independent.
- Normality: The data within each group should be approximately normally distributed.
- Homogeneity of variances: The variances of the groups should be approximately equal.
6. Calculate the overall mean: Compute the grand mean, which is the mean of all observations
across all groups.
10. Calculate the F-statistic: Compute the ratio of between-group mean square to within-group
mean square:
[ F = {MSB}/{MSW} ]
11. Determine the critical value or p-value: Based on the chosen significance level (α) and the
degrees of freedom, determine the critical value from the F-distribution table or use statistical
software to find the p-value associated with the F-statistic.
13. Interpret results: If the null hypothesis is rejected, conclude that there is evidence to suggest
that at least one group mean is different from the others.
14. Post-hoc tests (optional): If the ANOVA indicates significant differences between groups,
post-hoc tests (e.g., Tukey's HSD, Bonferroni correction, Scheffé's method) can be performed to
identify which specific group means differ from each other.
15. Conclusion: Provide a conclusion based on the results of the test, indicating whether there is
evidence to support the alternative hypothesis.
1. Formulate hypotheses:
- Null Hypothesis (H0): There is no significant difference in means across the conditions
or time points.
- Alternative Hypothesis (H1): At least one condition or time point has a different mean
from the others.
2. Set the significance level (α): Typically set at 0.05 or another predetermined value.
3. Collect data: Obtain data from subjects who are measured under different conditions
or at different time points. Each subject should provide multiple measurements.
4. Check assumptions:
- Sphericity: This assumption implies that the variances of the differences between all
possible pairs of conditions are equal.
- Normality: The differences between conditions should be approximately normally
distributed.
5. Calculate the within-subjects sum of squares (SSW): This measures the variability
between measurements within each subject.
6. Calculate the between-subjects sum of squares (SSB): This measures the variability
between subjects.
7. Calculate the total sum of squares (SST): This measures the total variability in the
data.
[ F = MSB/MSW ]
11. Determine the critical value or p-value: Based on the chosen significance level (α)
and the degrees of freedom, determine the critical value from the F-distribution table or
use statistical software to find the p-value associated with the F-statistic.
13. Interpret results: If the null hypothesis is rejected, conclude that there is evidence to
suggest that at least one condition or time point has a different mean from the others.
14. Conclusion: Provide a conclusion based on the results of the test, indicating whether
there is evidence to support the alternative hypothesis.
● RM-ANOVA assumes that the data meet the assumptions of sphericity and normality.
● Violations of these assumptions can affect the validity of the results.
● Additionally, RM-ANOVA is best suited for balanced designs, meaning equal numbers of
measurements for each subject in each condition or time point.
Mann-Whitney U test
● The Mann-Whitney U test, also known as the Mann-Whitney-Wilcoxon test or Wilcoxon
rank-sum test, is a non-parametric test used to compare two independent groups to
determine whether there is a significant difference between their distributions.
● It's commonly used when the assumptions of the t-test (a parametric test for comparing
means of two groups) are violated, such as when the data are not normally distributed or
when the variances are not equal.
1. Formulate hypotheses:
- Null Hypothesis (H0): There is no difference between the distributions of the two
groups.
- Alternative Hypothesis (H1): There is a difference between the distributions of the two
groups.
2. Set the significance level (α): Typically set at 0.05 or another predetermined value.
3. Collect data: Obtain data from two independent groups. Each group should have data
that represent the variable of interest (e.g., test scores, reaction times, etc.).
4. Rank the data: Pool the data from both groups and rank them from smallest to largest,
ignoring group membership. Ties are assigned the average of the ranks they would
occupy if un-tied.
5. Calculate the test statistic (U): The Mann-Whitney U test statistic (U) is calculated
based on the ranks of the data. It represents the sum of the ranks of one of the groups.
6. Determine the critical value or p-value: Based on the chosen significance level (α) and
the sample sizes of the groups, determine the critical value from the Mann-Whitney U
distribution table or use statistical software to find the p-value associated with the test
statistic.
9. Conclusion: Provide a conclusion based on the results of the test, indicating whether
there is evidence to support the alternative hypothesis.
● The Mann-Whitney U test is widely used in various fields, including psychology, biology,
and social sciences, as it provides a robust method for comparing distributions when
parametric assumptions cannot be met.
● The Wilcoxon signed-rank test, often referred to simply as the Wilcoxon test, is a
non-parametric statistical test used to determine whether there is a significant difference
between two related groups or conditions.
● It's specifically designed for paired data, where each observation in one group is paired
with a corresponding observation in the other group.
1. Formulate hypotheses:
- Null Hypothesis (H0): There is no difference between the two related groups or
conditions.
- Alternative Hypothesis (H1): There is a difference between the two related groups or
conditions.
2. Set the significance level (α): Typically set at 0.05 or another predetermined value.
3. Collect data: Obtain paired data from two related groups or conditions. Each pair
should represent measurements or observations taken under different conditions or at
different time points.
4. Calculate the differences: For each pair of observations, calculate the difference
between the two measurements.
5. Rank the absolute differences: Rank the absolute values of the differences from
smallest to largest, ignoring the signs.
6. Assign ranks: Assign ranks to the absolute differences based on their magnitude,
starting from 1 for the smallest absolute difference.
7. Calculate the test statistic (W): The Wilcoxon signed-rank test statistic (W) is the sum
of the ranks of the positive or negative differences, whichever is smaller.
8. Determine the critical value or p-value: Based on the chosen significance level (α) and
the sample size, determine the critical value from the Wilcoxon signed-rank distribution
table or use statistical software to find the p-value associated with the test statistic.
10. Interpret results: If the null hypothesis is rejected, conclude that there is evidence to
suggest that there is a significant difference between the two related groups or
conditions.
11. Conclusion: Provide a conclusion based on the results of the test, indicating whether
there is evidence to support the alternative hypothesis.
● The Wilcoxon signed-rank test is widely used in various fields, such as psychology,
medicine, and environmental science, when the data do not meet the assumptions of
parametric tests like the paired t-test.
● It provides a robust method for comparing paired data while allowing for flexibility and
sensitivity to deviations from normality.
Kruskal-Wallis test
● The Kruskal-Wallis test is a non-parametric test used to determine whether there are
statistically significant differences between the medians of three or more independent
groups.
● It's an extension of the Mann-Whitney U test for comparing two groups to situations with
more than two groups.
1. Formulate hypotheses:
- Null Hypothesis (H0): There is no difference between the medians of the groups.
- Alternative Hypothesis (H1): At least one group median is different from the others.
2. Set the significance level (α): Typically set at 0.05 or another predetermined value.
3. Collect data: Obtain data from three or more independent groups. Each group should have
data that represent the variable of interest (e.g., test scores, reaction times, etc.).
4. Rank the data: Pool the data from all groups and rank them from smallest to largest, ignoring
group membership. Ties are assigned the average of the ranks they would occupy if un-tied.
5. Calculate the test statistic (H): The Kruskal-Wallis test statistic (H) is calculated based on the
ranks of the data. It measures the degree of difference between the groups.
6. Determine the critical value or p-value: Based on the chosen significance level (α) and the
sample sizes of the groups, determine the critical value from the Kruskal-Wallis distribution table
or use statistical software to find the p-value associated with the test statistic.
8. Interpret results: If the null hypothesis is rejected, conclude that there is evidence to suggest
that at least one group median is different from the others.
9. Conclusion: Provide a conclusion based on the results of the test, indicating whether there is
evidence to support the alternative hypothesis.
● The Kruskal-Wallis test is particularly useful when the assumptions of parametric tests,
such as ANOVA, cannot be met, such as when the data are not normally distributed or
when the variances are not equal across groups.
● It's widely used in various fields, including psychology, biology, and social sciences, to
compare distributions between multiple groups.
1. Formulate hypotheses:
- Null Hypothesis (H0): There is no difference between the medians of the related groups.
- Alternative Hypothesis (H1): At least one group median is different from the others.
2. Set the significance level (α): Typically set at 0.05 or another predetermined value.
3. Collect data: Obtain data from three or more related groups or conditions. Each group should
have data that represent the variable of interest (e.g., test scores, reaction times, etc.).
4. Rank the data: For each subject or experimental unit, rank the observations within each
group from smallest to largest, ignoring ties.
5. Calculate the test statistic (Friedman's chi-squared statistic): Friedman's test statistic (χ²) is
calculated based on the ranks of the data. It measures the degree of difference between the
related groups.
6. Determine the critical value or p-value: Based on the chosen significance level (α) and the
sample size, determine the critical value from the chi-squared distribution table with \(k-1\)
degrees of freedom, where \(k\) is the number of related groups. Alternatively, use statistical
software to find the p-value associated with the test statistic.
8. Interpret results: If the null hypothesis is rejected, conclude that there is evidence to suggest
that at least one group median is different from the others.
9. Conclusion: Provide a conclusion based on the results of the test, indicating whether there is
evidence to support the alternative hypothesis.